Forráskód Böngészése

Remove non-bootstrap code

This commit removes steps for nodes that are legacy.

Not using bootstrapping config method is not supported in 3.10.
Michael Gugino 7 éve
szülő
commit
84896bf23b
27 módosított fájl, 4 hozzáadás és 266 törlés
  1. 0 1
      playbooks/aws/openshift-cluster/build_ami.yml
  2. 0 3
      playbooks/aws/openshift-cluster/install.yml
  3. 0 2
      playbooks/deploy_cluster.yml
  4. 0 7
      playbooks/gcp/openshift-cluster/build_image.yml
  5. 0 3
      playbooks/gcp/openshift-cluster/install.yml
  6. 0 13
      playbooks/init/evaluate_groups.yml
  7. 4 4
      playbooks/openshift-master/private/scaleup.yml
  8. 0 4
      playbooks/openshift-node/config.yml
  9. 0 69
      playbooks/openshift-node/private/additional_config.yml
  10. 0 42
      playbooks/openshift-node/private/config.yml
  11. 0 11
      playbooks/openshift-node/private/etcd_client_config.yml
  12. 0 13
      playbooks/openshift-node/private/manage_node.yml
  13. 0 1
      playbooks/openshift-node/scaleup.yml
  14. 0 4
      roles/openshift_aws/tasks/provision_instance.yml
  15. 0 3
      roles/openshift_gcp/tasks/setup_scale_group_facts.yml
  16. 0 2
      roles/openshift_node/defaults/main.yml
  17. 0 13
      roles/openshift_node/handlers/main.yml
  18. 0 2
      roles/openshift_node/tasks/aws.yml
  19. 0 54
      roles/openshift_node/tasks/config.yml
  20. 0 2
      roles/openshift_node/tasks/config/configure-node-settings.yml
  21. 0 2
      roles/openshift_node/tasks/config/configure-proxy-settings.yml
  22. 0 3
      roles/openshift_node/tasks/dnsmasq/no-network-manager.yml
  23. 0 1
      roles/openshift_node/tasks/main.yml
  24. 0 4
      roles/openshift_node/tasks/registry_auth.yml
  25. 0 1
      roles/openshift_node/tasks/systemd_units.yml
  26. 0 1
      roles/openshift_node/tasks/upgrade.yml
  27. 0 1
      roles/openshift_node/tasks/upgrade/config_changes.yml

+ 0 - 1
playbooks/aws/openshift-cluster/build_ami.yml

@@ -27,7 +27,6 @@
   - name: set the user to perform installation
     set_fact:
       ansible_ssh_user: "{{ openshift_aws_build_ami_ssh_user | default(ansible_ssh_user) }}"
-      openshift_node_bootstrap: True
       openshift_node_image_prep_packages:
       - cloud-utils-growpart
 

+ 0 - 3
playbooks/aws/openshift-cluster/install.yml

@@ -20,6 +20,3 @@
 
 - name: configure the control plane
   import_playbook: ../../common/private/control_plane.yml
-
-- name: ensure the masters are configured as nodes
-  import_playbook: ../../openshift-node/private/config.yml

+ 0 - 2
playbooks/deploy_cluster.yml

@@ -7,6 +7,4 @@
 
 - import_playbook: openshift-node/private/join.yml
 
-- import_playbook: openshift-node/private/config.yml
-
 - import_playbook: common/private/components.yml

+ 0 - 7
playbooks/gcp/openshift-cluster/build_image.yml

@@ -26,7 +26,6 @@
   tasks:
   - name: Set facts
     set_fact:
-      openshift_node_bootstrap: True
       openshift_master_unsupported_embedded_etcd: True
 
   - name: Create the image instance disk
@@ -78,12 +77,6 @@
   tasks:
   - wait_for_connection:
 
-- hosts: nodes
-  tasks:
-  - name: Set facts
-    set_fact:
-      openshift_node_bootstrap: True
-
 # This is the part that installs all of the software and configs for the instance
 # to become a node.
 - import_playbook: ../../openshift-node/private/image_prep.yml

+ 0 - 3
playbooks/gcp/openshift-cluster/install.yml

@@ -20,9 +20,6 @@
 - name: run the GCP specific post steps
   import_playbook: install_gcp.yml
 
-- name: configure any nodes that aren't bootstrapped
-  import_playbook: ../../openshift-node/private/config.yml
-
 - name: install components
   import_playbook: ../../common/private/components.yml
 

+ 0 - 13
playbooks/init/evaluate_groups.yml

@@ -42,18 +42,6 @@
       msg: This playbook requires g_glusterfs_hosts to be set
     when: g_glusterfs_hosts is not defined
 
-  - name: Evaluate groups - Fail if no etcd hosts group is defined
-    fail:
-      msg: >
-        Running etcd as an embedded service is no longer supported. If this is a
-        new install please define an 'etcd' group with either one, three or five
-        hosts. These hosts may be the same hosts as your masters. If this is an
-        upgrade please see https://docs.openshift.com/container-platform/latest/install_config/upgrading/migrating_embedded_etcd.html
-        for documentation on how to migrate from embedded to external etcd.
-    when:
-    - g_etcd_hosts | default([]) | length == 0
-    - not (openshift_node_bootstrap | default(True))
-
   - name: Evaluate oo_all_hosts
     add_host:
       name: "{{ item }}"
@@ -151,7 +139,6 @@
       ansible_become: "{{ g_sudo | default(omit) }}"
     with_items: "{{ g_new_node_hosts | default(g_node_hosts | default([], true), true) }}"
     changed_when: no
-    when: hostvars[item].openshift_node_bootstrap | default(True) | bool
 
   - name: Add masters to oo_nodes_to_bootstrap
     add_host:

+ 4 - 4
playbooks/openshift-master/private/scaleup.yml

@@ -36,12 +36,12 @@
 
 - import_playbook: set_network_facts.yml
 
+- import_playbook: ../../openshift-node/private/bootstrap.yml
+
 - import_playbook: ../../openshift-etcd/private/master_etcd_certificates.yml
 
 - import_playbook: config.yml
 
-- import_playbook: ../../openshift-loadbalancer/private/config.yml
-
-- import_playbook: ../../openshift-node/private/certificates.yml
+- import_playbook: ../../openshift-node/private/join.yml
 
-- import_playbook: ../../openshift-node/private/config.yml
+- import_playbook: ../../openshift-loadbalancer/private/config.yml

+ 0 - 4
playbooks/openshift-node/config.yml

@@ -1,4 +0,0 @@
----
-- import_playbook: ../init/main.yml
-
-- import_playbook: private/config.yml

+ 0 - 69
playbooks/openshift-node/private/additional_config.yml

@@ -1,69 +0,0 @@
----
-- name: create additional node network plugin groups
-  hosts: "{{ openshift_node_scale_up_group | default('oo_nodes_to_config') }}:!oo_nodes_to_bootstrap"
-  tasks:
-  # Creating these node groups will prevent a ton of skipped tasks.
-  # Create group for flannel nodes
-  - group_by:
-      key: oo_nodes_use_{{ (openshift_use_flannel | default(False)) | ternary('flannel','nothing') }}
-    changed_when: False
-  # Create group for calico nodes
-  - group_by:
-      key: oo_nodes_use_{{ (openshift_use_calico | default(False)) | ternary('calico','nothing') }}
-    changed_when: False
-  # Create group for nuage nodes
-  - group_by:
-      key: oo_nodes_use_{{ (openshift_use_nuage | default(False)) | ternary('nuage','nothing') }}
-    changed_when: False
-  # Create group for contiv nodes
-  - group_by:
-      key: oo_nodes_use_{{ (openshift_use_contiv | default(False)) | ternary('contiv','nothing') }}
-    changed_when: False
-  # Create group for kuryr nodes
-  - group_by:
-      key: oo_nodes_use_{{ (openshift_use_kuryr | default(False)) | ternary('kuryr','nothing') }}
-    changed_when: False
-
-- import_playbook: etcd_client_config.yml
-  vars:
-    openshift_node_scale_up_group: "oo_nodes_use_flannel:oo_nodes_use_calico:oo_nodes_use_contiv:oo_nodes_use_kuryr"
-
-- name: Additional node config
-  hosts: oo_nodes_use_flannel
-  roles:
-  - role: flannel
-    etcd_urls: "{{ hostvars[groups.oo_first_master.0].openshift_master_etcd_urls }}"
-    when: openshift_use_flannel | default(false) | bool
-
-- name: Additional node config
-  hosts: oo_nodes_use_calico
-  roles:
-  - role: calico
-    when: openshift_use_calico | default(false) | bool
-
-- name: Additional node config
-  hosts: oo_nodes_use_nuage
-  roles:
-  - role: nuage_node
-    when: openshift_use_nuage | default(false) | bool
-
-- name: Configure Contiv masters
-  hosts: oo_masters_to_config
-  roles:
-  - role: contiv
-    contiv_master: true
-    when: openshift_use_contiv | default(false) | bool
-
-- name: Configure rest of Contiv nodes
-  hosts: "{{ groups.oo_nodes_use_contiv | default([]) | difference(groups.oo_masters_to_config) }}"
-  roles:
-  - role: contiv
-    when: openshift_use_contiv | default(false) | bool
-
-- name: Configure Kuryr node
-  hosts: oo_nodes_use_kuryr
-  tasks:
-  - import_role:
-      name: kuryr
-      tasks_from: node
-    when: openshift_use_kuryr | default(false) | bool

+ 0 - 42
playbooks/openshift-node/private/config.yml

@@ -1,42 +0,0 @@
----
-- name: Node Install Checkpoint Start
-  hosts: all
-  gather_facts: false
-  tasks:
-  - name: Set Node install 'In Progress'
-    run_once: true
-    set_stats:
-      data:
-        installer_phase_node:
-          status: "In Progress"
-          start: "{{ lookup('pipe', 'date +%Y%m%d%H%M%SZ') }}"
-
-- import_playbook: disable_excluders.yml
-  vars:
-    l_node_group: oo_nodes_to_config:!oo_nodes_to_bootstrap
-
-- import_playbook: certificates.yml
-
-- import_playbook: configure_nodes.yml
-  vars:
-    l_node_group: oo_nodes_to_config:!oo_nodes_to_bootstrap
-
-- import_playbook: additional_config.yml
-
-- import_playbook: manage_node.yml
-
-- import_playbook: enable_excluders.yml
-  vars:
-    l_node_group: oo_nodes_to_config:!oo_nodes_to_bootstrap
-
-- name: Node Install Checkpoint End
-  hosts: all
-  gather_facts: false
-  tasks:
-  - name: Set Node install 'Complete'
-    run_once: true
-    set_stats:
-      data:
-        installer_phase_node:
-          status: "Complete"
-          end: "{{ lookup('pipe', 'date +%Y%m%d%H%M%SZ') }}"

+ 0 - 11
playbooks/openshift-node/private/etcd_client_config.yml

@@ -1,11 +0,0 @@
----
-- name: etcd_client node config
-  hosts: "{{ openshift_node_scale_up_group | default('this_group_does_not_exist') }}"
-  tasks:
-  - import_role:
-      name: etcd
-      tasks_from: client_certificates.yml
-    vars:
-      etcd_cert_prefix: flannel.etcd-
-      etcd_cert_subdir: "openshift-node-{{ openshift.common.hostname }}"
-      etcd_cert_config_dir: "{{ openshift.common.config_base }}/node"

+ 0 - 13
playbooks/openshift-node/private/manage_node.yml

@@ -1,13 +0,0 @@
----
-- name: Additional node config
-  hosts: "{{ openshift_node_scale_up_group | default('oo_nodes_to_config') }}:!oo_nodes_to_bootstrap"
-  vars:
-    openshift_node_master_api_url: "{{ hostvars[groups.oo_first_master.0].openshift.master.api_url }}"
-  roles:
-  - role: openshift_manage_node
-    openshift_master_host: "{{ groups.oo_first_master.0 }}"
-    openshift_manage_node_is_master: "{{ ('oo_masters_to_config' in group_names) | bool }}"
-  tasks:
-  - name: Create group for deployment type
-    group_by: key=oo_nodes_deployment_type_{{ openshift_deployment_type }}
-    changed_when: False

+ 0 - 1
playbooks/openshift-node/scaleup.yml

@@ -37,4 +37,3 @@
 
 - import_playbook: private/bootstrap.yml
 - import_playbook: private/join.yml
-- import_playbook: private/config.yml

+ 0 - 4
roles/openshift_aws/tasks/provision_instance.yml

@@ -1,8 +1,4 @@
 ---
-- name: set openshift_node_bootstrap to True when building AMI
-  set_fact:
-    openshift_node_bootstrap: True
-
 - include_tasks: vpc_and_subnet_id.yml
 
 - name: create instance for ami creation

+ 0 - 3
roles/openshift_gcp/tasks/setup_scale_group_facts.yml

@@ -31,14 +31,12 @@
   add_host:
     name: "{{ hostvars[item].gce_name }}"
     groups: nodes, new_nodes
-    openshift_node_bootstrap: "{{ openshift_node_bootstrap | default(True) }}"
   with_items: "{{ groups['tag_ocp-node'] | default([]) | difference(groups['tag_ocp-bootstrap'] | default([])) }}"
 
 - name: Add bootstrap node instances
   add_host:
     name: "{{ hostvars[item].gce_name }}"
     groups: bootstrap_nodes
-    openshift_node_bootstrap: True
     openshift_is_bootstrapped: True
   with_items: "{{ groups['tag_ocp-node'] | default([]) | intersect(groups['tag_ocp-bootstrap'] | default([])) }}"
 
@@ -46,7 +44,6 @@
   add_host:
     name: "{{ item }}"
     groups: nodes, new_nodes
-    openshift_node_bootstrap: True
     openshift_is_bootstrapped: True
   with_items: "{{ groups['tag_ocp-bootstrap'] | default([]) }}"
   when: all_nodes | default(False)

+ 0 - 2
roles/openshift_node/defaults/main.yml

@@ -144,8 +144,6 @@ default_r_openshift_node_image_prep_packages:
 - ceph-common
 r_openshift_node_image_prep_packages: "{{ default_r_openshift_node_image_prep_packages | union(openshift_node_image_prep_packages | default([])) }}"
 
-openshift_node_bootstrap: True
-
 r_openshift_node_os_firewall_deny: []
 default_r_openshift_node_os_firewall_allow:
 - service: Kubernetes kubelet

+ 0 - 13
roles/openshift_node/handlers/main.yml

@@ -14,19 +14,6 @@
   when:
   - (not skip_node_svc_handlers | default(False) | bool)
 
-- name: restart node
-  systemd:
-    name: "{{ openshift_service_type }}-node"
-    state: restarted
-  register: l_openshift_node_restart_node_result
-  until: not (l_openshift_node_restart_node_result is failed)
-  retries: 3
-  delay: 30
-  when:
-  - (not skip_node_svc_handlers | default(False) | bool)
-  - not (node_service_status_changed | default(false) | bool)
-  - not openshift_node_bootstrap
-
 - name: reload systemd units
   command: systemctl daemon-reload
   when:

+ 0 - 2
roles/openshift_node/tasks/aws.yml

@@ -17,5 +17,3 @@
     - openshift_cloudprovider_kind == 'aws'
     - openshift_cloudprovider_aws_access_key is defined
     - openshift_cloudprovider_aws_secret_key is defined
-  notify:
-    - restart node

+ 0 - 54
roles/openshift_node/tasks/config.yml

@@ -20,8 +20,6 @@
     owner: root
     group: root
     mode: 0600
-  notify:
-    - restart node
 
 - name: Configure Node Environment Variables
   lineinfile:
@@ -30,8 +28,6 @@
     line: "{{ item.key }}={{ item.value }}"
     create: true
   with_dict: "{{ openshift_node_env_vars }}"
-  notify:
-    - restart node
 
 - name: Ensure the node static pod directory exists
   file:
@@ -42,53 +38,3 @@
 - name: include aws provider credentials
   import_tasks: aws.yml
   when: not (openshift_node_use_instance_profiles | default(False))
-
-# Necessary because when you're on a node that's also a master the master will be
-# restarted after the node restarts docker and it will take up to 60 seconds for
-# systemd to start the master again
-- when:
-    - openshift_is_containerized | bool
-    - not openshift_node_bootstrap
-  block:
-    - name: Wait for master API to become available before proceeding
-      # Using curl here since the uri module requires python-httplib2 and
-      # wait_for port doesn't provide health information.
-      command: >
-        curl --silent --tlsv1.2 --max-time 2
-        --cacert {{ openshift.common.config_base }}/node/ca.crt
-        {{ openshift_node_master_api_url }}/healthz/ready
-      args:
-        # Disables the following warning:
-        # Consider using get_url or uri module rather than running curl
-        warn: no
-      register: api_available_output
-      until: api_available_output.stdout == 'ok'
-      retries: 120
-      delay: 1
-      changed_when: false
-
-- when: not openshift_node_bootstrap
-  block:
-    - name: Start and enable node
-      systemd:
-        name: "{{ openshift_service_type }}-node"
-        enabled: yes
-        state: started
-        daemon_reload: yes
-      register: node_start_result
-      until: not node_start_result is failed
-      retries: 1
-      delay: 30
-      ignore_errors: true
-
-    - name: Dump logs from node service if it failed
-      command: journalctl --no-pager -n 100 -u {{ openshift_service_type }}-node
-      when: node_start_result is failed
-
-    - name: Abort if node failed to start
-      fail:
-        msg: Node failed to start please inspect the logs and try again
-      when: node_start_result is failed
-
-    - set_fact:
-        node_service_status_changed: "{{ node_start_result is changed }}"

+ 0 - 2
roles/openshift_node/tasks/config/configure-node-settings.yml

@@ -14,5 +14,3 @@
     line: "CONFIG_FILE={{ openshift.common.config_base }}/node/node-config.yaml"
   - regex: '^IMAGE_VERSION='
     line: "IMAGE_VERSION={{ openshift_image_tag }}"
-  notify:
-  - restart node

+ 0 - 2
roles/openshift_node/tasks/config/configure-proxy-settings.yml

@@ -13,5 +13,3 @@
   - regex: '^NO_PROXY='
     line: "NO_PROXY={{ openshift.common.no_proxy | default([]) }},{{ openshift.common.portal_net }},{{ hostvars[groups.oo_first_master.0].openshift.master.sdn_cluster_network_cidr }}"
   when: ('http_proxy' in openshift.common and openshift.common.http_proxy != '')
-  notify:
-  - restart node

+ 0 - 3
roles/openshift_node/tasks/dnsmasq/no-network-manager.yml

@@ -1,7 +1,4 @@
 ---
-- fail: msg="Currently, NetworkManager must be installed and enabled prior to installation."
-  when: not openshift_node_bootstrap | bool
-
 - name: Install NetworkManager during node_bootstrap provisioning
   package:
     name: NetworkManager

+ 0 - 1
roles/openshift_node/tasks/main.yml

@@ -42,7 +42,6 @@
     name: NetworkManager
     enabled: yes
     state: restarted
-  when: openshift_node_bootstrap | bool
 
 # The atomic-openshift-node service will set this parameter on
 # startup, but if the network service is restarted this setting is

+ 0 - 4
roles/openshift_node/tasks/registry_auth.yml

@@ -15,8 +15,6 @@
   retries: 3
   delay: 5
   until: node_oreg_auth_credentials_create.rc == 0
-  notify:
-    - restart node
 
 # docker_creds is a custom module from lib_utils
 # 'docker login' requires a docker.service running on the local host, this is an
@@ -33,8 +31,6 @@
     - oreg_auth_user is defined
     - (not node_oreg_auth_credentials_stat.stat.exists or oreg_auth_credentials_replace) | bool
   register: node_oreg_auth_credentials_create_alt
-  notify:
-    - restart node
 
 # Container images may need the registry credentials
 - name: Setup ro mount of /root/.docker for containerized hosts

+ 0 - 1
roles/openshift_node/tasks/systemd_units.yml

@@ -11,7 +11,6 @@
   when: not l_is_node_system_container | bool
   notify:
   - reload systemd units
-  - restart node
 
 - import_tasks: config/configure-node-settings.yml
 - import_tasks: config/configure-proxy-settings.yml

+ 0 - 1
roles/openshift_node/tasks/upgrade.yml

@@ -43,7 +43,6 @@
     fail_on_timeout: true
   delegate_to: "{{ groups.oo_first_master.0 }}"
   ignore_errors: true
-  when: openshift_node_bootstrap | default(True) | bool
 
 - name: Wait for node to be ready
   oc_obj:

+ 0 - 1
roles/openshift_node/tasks/upgrade/config_changes.yml

@@ -57,7 +57,6 @@
 
 - name: Move existing credentials and configuration into bootstrap configuration
   import_tasks: bootstrap_changes.yml
-  when: openshift_node_bootstrap | default(True) | bool
 
 - name: Reset selinux context
   command: restorecon -RF {{ openshift_node_data_dir }}/openshift.local.volumes