diff options
Diffstat (limited to 'playbooks/common')
35 files changed, 47 insertions, 892 deletions
| diff --git a/playbooks/common/openshift-cluster/cockpit-ui.yml b/playbooks/common/openshift-cluster/cockpit-ui.yml index 5ddafdb07..359132dd0 100644 --- a/playbooks/common/openshift-cluster/cockpit-ui.yml +++ b/playbooks/common/openshift-cluster/cockpit-ui.yml @@ -3,4 +3,6 @@    hosts: oo_first_master    roles:    - role: cockpit-ui -    when: ( openshift.common.version_gte_3_3_or_1_3  | bool ) and ( openshift_hosted_manage_registry | default(true) | bool ) and not (openshift.docker.hosted_registry_insecure | default(false) | bool) +    when: +    - openshift_hosted_manage_registry | default(true) | bool +    - not openshift.docker.hosted_registry_insecure | default(false) | bool diff --git a/playbooks/common/openshift-cluster/enable_dnsmasq.yml b/playbooks/common/openshift-cluster/enable_dnsmasq.yml index be14b06f0..f91361b67 100644 --- a/playbooks/common/openshift-cluster/enable_dnsmasq.yml +++ b/playbooks/common/openshift-cluster/enable_dnsmasq.yml @@ -1,13 +1,8 @@  --- -- include: evaluate_groups.yml -  - name: Load openshift_facts    hosts: oo_masters_to_config:oo_nodes_to_config    roles:    - openshift_facts -  post_tasks: -  - fail: msg="This playbook requires a master version of at least Origin 1.1 or OSE 3.1" -    when: not openshift.common.version_gte_3_1_1_or_1_1_1 | bool  - name: Reconfigure masters to listen on our new dns_port    hosts: oo_masters_to_config diff --git a/playbooks/common/openshift-cluster/evaluate_groups.yml b/playbooks/common/openshift-cluster/evaluate_groups.yml deleted file mode 100644 index 78b552279..000000000 --- a/playbooks/common/openshift-cluster/evaluate_groups.yml +++ /dev/null @@ -1,190 +0,0 @@ ---- -- name: Populate config host groups -  hosts: localhost -  connection: local -  become: no -  gather_facts: no -  tasks: -  - name: Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required -    fail: -      msg: This playbook requires g_etcd_hosts or g_new_etcd_hosts to be set -    when: g_etcd_hosts is not defined and g_new_etcd_hosts is not defined - -  - name: Evaluate groups - g_master_hosts or g_new_master_hosts required -    fail: -      msg: This playbook requires g_master_hosts or g_new_master_hosts to be set -    when: g_master_hosts is not defined and g_new_master_hosts is not defined - -  - name: Evaluate groups - g_node_hosts or g_new_node_hosts required -    fail: -      msg: This playbook requires g_node_hosts or g_new_node_hosts to be set -    when: g_node_hosts is not defined and g_new_node_hosts is not defined - -  - name: Evaluate groups - g_lb_hosts required -    fail: -      msg: This playbook requires g_lb_hosts to be set -    when: g_lb_hosts is not defined - -  - name: Evaluate groups - g_nfs_hosts required -    fail: -      msg: This playbook requires g_nfs_hosts to be set -    when: g_nfs_hosts is not defined - -  - name: Evaluate groups - g_nfs_hosts is single host -    fail: -      msg: The nfs group must be limited to one host -    when: g_nfs_hosts | default([]) | length > 1 - -  - name: Evaluate groups - g_glusterfs_hosts required -    fail: -      msg: This playbook requires g_glusterfs_hosts to be set -    when: g_glusterfs_hosts is not defined - -  - name: Evaluate groups - Fail if no etcd hosts group is defined -    fail: -      msg: > -        Running etcd as an embedded service is no longer supported. If this is a -        new install please define an 'etcd' group with either one or three -        hosts. These hosts may be the same hosts as your masters. If this is an -        upgrade you may set openshift_master_unsupported_embedded_etcd=true -        until a migration playbook becomes available. -    when: -    - g_etcd_hosts | default([]) | length not in [3,1] -    - not openshift_master_unsupported_embedded_etcd | default(False) -    - not (openshift_node_bootstrap | default(False)) - -  - name: Evaluate oo_all_hosts -    add_host: -      name: "{{ item }}" -      groups: oo_all_hosts -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_all_hosts | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_masters -    add_host: -      name: "{{ item }}" -      groups: oo_masters -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_master_hosts | union(g_new_master_hosts) | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_first_master -    add_host: -      name: "{{ g_master_hosts[0] }}" -      groups: oo_first_master -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    when: g_master_hosts|length > 0 -    changed_when: no - -  - name: Evaluate oo_new_etcd_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_new_etcd_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_new_etcd_hosts | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_masters_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_masters_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_new_master_hosts | default(g_master_hosts | default([], true), true) }}" -    changed_when: no - -  - name: Evaluate oo_etcd_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_etcd_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_etcd_hosts | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_first_etcd -    add_host: -      name: "{{ g_etcd_hosts[0] }}" -      groups: oo_first_etcd -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    when: g_etcd_hosts|length > 0 -    changed_when: no - -  # We use two groups one for hosts we're upgrading which doesn't include embedded etcd -  # The other for backing up which includes the embedded etcd host, there's no need to -  # upgrade embedded etcd that just happens when the master is updated. -  - name: Evaluate oo_etcd_hosts_to_upgrade -    add_host: -      name: "{{ item }}" -      groups: oo_etcd_hosts_to_upgrade -    with_items: "{{ groups.oo_etcd_to_config if groups.oo_etcd_to_config is defined and groups.oo_etcd_to_config | length > 0 else [] }}" -    changed_when: False - -  - name: Evaluate oo_etcd_hosts_to_backup -    add_host: -      name: "{{ item }}" -      groups: oo_etcd_hosts_to_backup -    with_items: "{{ groups.oo_etcd_to_config if groups.oo_etcd_to_config is defined and groups.oo_etcd_to_config | length > 0 else (groups.oo_first_master | default([])) }}" -    changed_when: False - -  - name: Evaluate oo_nodes_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_nodes_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_new_node_hosts | default(g_node_hosts | default([], true), true) }}" -    changed_when: no - -  # Skip adding the master to oo_nodes_to_config when g_new_node_hosts is -  - name: Add master to oo_nodes_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_nodes_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_master_hosts | default([]) }}" -    when: g_nodeonmaster | default(false) | bool and not g_new_node_hosts | default(false) | bool -    changed_when: no - -  - name: Evaluate oo_lb_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_lb_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_lb_hosts | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_nfs_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_nfs_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_nfs_hosts | default([]) }}" -    changed_when: no - -  - name: Evaluate oo_glusterfs_to_config -    add_host: -      name: "{{ item }}" -      groups: oo_glusterfs_to_config -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ g_glusterfs_hosts | union(g_glusterfs_registry_hosts | default([])) }}" -    changed_when: no - -  - name: Evaluate oo_etcd_to_migrate -    add_host: -      name: "{{ item }}" -      groups: oo_etcd_to_migrate -      ansible_ssh_user: "{{ g_ssh_user | default(omit) }}" -      ansible_become: "{{ g_sudo | default(omit) }}" -    with_items: "{{ groups.oo_etcd_to_config if groups.oo_etcd_to_config | default([]) | length != 0 else (groups.oo_first_master |default([]))}}" -    changed_when: no diff --git a/playbooks/common/openshift-cluster/initialize_facts.yml b/playbooks/common/openshift-cluster/initialize_facts.yml deleted file mode 100644 index 91223d368..000000000 --- a/playbooks/common/openshift-cluster/initialize_facts.yml +++ /dev/null @@ -1,169 +0,0 @@ ---- -- name: Ensure that all non-node hosts are accessible -  hosts: oo_masters_to_config:oo_etcd_to_config:oo_lb_to_config:oo_nfs_to_config -  any_errors_fatal: true -  tasks: - -- name: Initialize host facts -  hosts: oo_all_hosts -  tasks: -  - name: load openshift_facts module -    include_role: -      name: openshift_facts -    static: yes - -  # TODO: Should this role be refactored into health_checks?? -  - name: Run openshift_sanitize_inventory to set variables -    include_role: -      name: openshift_sanitize_inventory - -  - name: Detecting Operating System from ostree_booted -    stat: -      path: /run/ostree-booted -    register: ostree_booted - -  # Locally setup containerized facts for now -  - name: initialize_facts set fact l_is_atomic -    set_fact: -      l_is_atomic: "{{ ostree_booted.stat.exists }}" - -  - name: initialize_facts set fact for containerized and l_is_*_system_container -    set_fact: -      l_is_containerized: "{{ (l_is_atomic | bool) or (containerized | default(false) | bool) }}" -      l_is_openvswitch_system_container: "{{ (openshift_use_openvswitch_system_container | default(openshift_use_system_containers | default(false)) | bool) }}" -      l_is_node_system_container: "{{ (openshift_use_node_system_container | default(openshift_use_system_containers | default(false)) | bool) }}" -      l_is_master_system_container: "{{ (openshift_use_master_system_container | default(openshift_use_system_containers | default(false)) | bool) }}" -      l_is_etcd_system_container: "{{ (openshift_use_etcd_system_container | default(openshift_use_system_containers | default(false)) | bool) }}" - -  - name: initialize_facts set facts for l_any_system_container -    set_fact: -      l_any_system_container: "{{ l_is_etcd_system_container or l_is_openvswitch_system_container or l_is_node_system_container or l_is_master_system_container }}" - -  - name: initialize_facts set fact for l_etcd_runtime -    set_fact: -      l_etcd_runtime: "{{ 'runc' if l_is_etcd_system_container else 'docker' if l_is_containerized else 'host' }}" - -  # TODO: Should this be moved into health checks?? -  # Seems as though any check that happens with a corresponding fail should move into health_checks -  - name: Validate python version - ans_dist is fedora and python is v3 -    fail: -      msg: | -        openshift-ansible requires Python 3 for {{ ansible_distribution }}; -        For information on enabling Python 3 with Ansible, see https://docs.ansible.com/ansible/python_3_support.html -    when: -    - ansible_distribution == 'Fedora' -    - ansible_python['version']['major'] != 3 - -  # TODO: Should this be moved into health checks?? -  # Seems as though any check that happens with a corresponding fail should move into health_checks -  - name: Validate python version - ans_dist not Fedora and python must be v2 -    fail: -      msg: "openshift-ansible requires Python 2 for {{ ansible_distribution }}" -    when: -    - ansible_distribution != 'Fedora' -    - ansible_python['version']['major'] != 2 - -  # TODO: Should this be moved into health checks?? -  # Seems as though any check that happens with a corresponding fail should move into health_checks -  # Fail as early as possible if Atomic and old version of Docker -  - when: -    - l_is_atomic | bool -    block: - -    # See https://access.redhat.com/articles/2317361 -    # and https://github.com/ansible/ansible/issues/15892 -    # NOTE: the "'s can not be removed at this level else the docker command will fail -    # NOTE: When ansible >2.2.1.x is used this can be updated per -    # https://github.com/openshift/openshift-ansible/pull/3475#discussion_r103525121 -    - name: Determine Atomic Host Docker Version -      shell: 'CURLY="{"; docker version --format "$CURLY{json .Server.Version}}"' -      register: l_atomic_docker_version - -    - name: assert atomic host docker version is 1.12 or later -      assert: -        that: -        - l_atomic_docker_version.stdout | replace('"', '') | version_compare('1.12','>=') -        msg: Installation on Atomic Host requires Docker 1.12 or later. Please upgrade and restart the Atomic Host. - -  - when: -    - not l_is_atomic | bool -    block: -    - name: Ensure openshift-ansible installer package deps are installed -      package: -        name: "{{ item }}" -        state: present -      with_items: -      - iproute -      - "{{ 'python3-dbus' if ansible_distribution == 'Fedora' else 'dbus-python' }}" -      - "{{ 'python3-PyYAML' if ansible_distribution == 'Fedora' else 'PyYAML' }}" -      - yum-utils - -    - name: Ensure various deps for running system containers are installed -      package: -        name: "{{ item }}" -        state: present -      with_items: -      - atomic -      - ostree -      - runc -      when: -      - l_any_system_container | bool - -  - name: Default system_images_registry to a enterprise registry -    set_fact: -      system_images_registry: "registry.access.redhat.com" -    when: -    - system_images_registry is not defined -    - openshift_deployment_type == "openshift-enterprise" - -  - name: Default system_images_registry to community registry -    set_fact: -      system_images_registry: "docker.io" -    when: -    - system_images_registry is not defined -    - openshift_deployment_type == "origin" - -  - name: Gather Cluster facts and set is_containerized if needed -    openshift_facts: -      role: common -      local_facts: -        deployment_type: "{{ openshift_deployment_type }}" -        deployment_subtype: "{{ openshift_deployment_subtype | default(None) }}" -        cli_image: "{{ osm_image | default(None) }}" -        hostname: "{{ openshift_hostname | default(None) }}" -        ip: "{{ openshift_ip | default(None) }}" -        is_containerized: "{{ l_is_containerized | default(None) }}" -        is_openvswitch_system_container: "{{ l_is_openvswitch_system_container | default(false) }}" -        is_node_system_container: "{{ l_is_node_system_container | default(false) }}" -        is_master_system_container: "{{ l_is_master_system_container | default(false) }}" -        is_etcd_system_container: "{{ l_is_etcd_system_container | default(false) }}" -        etcd_runtime: "{{ l_etcd_runtime }}" -        system_images_registry: "{{ system_images_registry }}" -        public_hostname: "{{ openshift_public_hostname | default(None) }}" -        public_ip: "{{ openshift_public_ip | default(None) }}" -        portal_net: "{{ openshift_portal_net | default(openshift_master_portal_net) | default(None) }}" -        http_proxy: "{{ openshift_http_proxy | default(None) }}" -        https_proxy: "{{ openshift_https_proxy | default(None) }}" -        no_proxy: "{{ openshift_no_proxy | default(None) }}" -        generate_no_proxy_hosts: "{{ openshift_generate_no_proxy_hosts | default(True) }}" - -  - name: Set fact of no_proxy_internal_hostnames -    openshift_facts: -      role: common -      local_facts: -        no_proxy_internal_hostnames: "{{ hostvars | oo_select_keys(groups['oo_nodes_to_config'] -                                             | union(groups['oo_masters_to_config']) -                                             | union(groups['oo_etcd_to_config'] | default([]))) -                                         | oo_collect('openshift.common.hostname') | default([]) | join (',') -                                         }}" -    when: -    - openshift_http_proxy is defined or openshift_https_proxy is defined -    - openshift_generate_no_proxy_hosts | default(True) | bool - -  - name: initialize_facts set_fact repoquery command -    set_fact: -      repoquery_cmd: "{{ 'dnf repoquery --latest-limit 1 -d 0' if ansible_pkg_mgr == 'dnf' else 'repoquery --plugins' }}" - -  - name: initialize_facts set_fact on openshift_docker_hosted_registry_network -    set_fact: -      openshift_docker_hosted_registry_network: "{{ '' if 'oo_first_master' not in groups else hostvars[groups.oo_first_master.0].openshift.common.portal_net }}" diff --git a/playbooks/common/openshift-cluster/initialize_openshift_repos.yml b/playbooks/common/openshift-cluster/initialize_openshift_repos.yml deleted file mode 100644 index a7114fc80..000000000 --- a/playbooks/common/openshift-cluster/initialize_openshift_repos.yml +++ /dev/null @@ -1,8 +0,0 @@ ---- -- name: Setup yum repositories for all hosts -  hosts: oo_all_hosts -  gather_facts: no -  tasks: -  - name: initialize openshift repos -    include_role: -      name: openshift_repos diff --git a/playbooks/common/openshift-cluster/initialize_openshift_version.yml b/playbooks/common/openshift-cluster/initialize_openshift_version.yml deleted file mode 100644 index 37a5284d5..000000000 --- a/playbooks/common/openshift-cluster/initialize_openshift_version.yml +++ /dev/null @@ -1,21 +0,0 @@ ---- -# NOTE: requires openshift_facts be run -- name: Determine openshift_version to configure on first master -  hosts: oo_first_master -  roles: -  - openshift_version - -# NOTE: We set this even on etcd hosts as they may also later run as masters, -# and we don't want to install wrong version of docker and have to downgrade -# later. -- name: Set openshift_version for etcd, node, and master hosts -  hosts: oo_etcd_to_config:oo_nodes_to_config:oo_masters_to_config:!oo_first_master -  vars: -    openshift_version: "{{ hostvars[groups.oo_first_master.0].openshift_version }}" -  pre_tasks: -  - set_fact: -      openshift_pkg_version: -{{ openshift_version }} -    when: openshift_pkg_version is not defined -  - debug: msg="openshift_pkg_version set to {{ openshift_pkg_version }}" -  roles: -  - openshift_version diff --git a/playbooks/common/openshift-cluster/redeploy-certificates/openshift-ca.yml b/playbooks/common/openshift-cluster/redeploy-certificates/openshift-ca.yml index e22c8cbdb..eb225dfb5 100644 --- a/playbooks/common/openshift-cluster/redeploy-certificates/openshift-ca.yml +++ b/playbooks/common/openshift-cluster/redeploy-certificates/openshift-ca.yml @@ -1,11 +1,4 @@  --- -- name: Verify OpenShift version is greater than or equal to 1.2 or 3.2 -  hosts: oo_first_master -  tasks: -  - fail: -      msg: "The current OpenShift version is less than 1.2/3.2 and does not support CA bundles." -    when: not openshift.common.version_gte_3_2_or_1_2 | bool -  - name: Check cert expirys    hosts: oo_nodes_to_config:oo_masters_to_config:oo_etcd_to_config    vars: diff --git a/playbooks/common/openshift-cluster/redeploy-certificates/registry.yml b/playbooks/common/openshift-cluster/redeploy-certificates/registry.yml index afd5463b2..7e9363c5f 100644 --- a/playbooks/common/openshift-cluster/redeploy-certificates/registry.yml +++ b/playbooks/common/openshift-cluster/redeploy-certificates/registry.yml @@ -70,9 +70,7 @@          --hostnames="{{ docker_registry_service_ip.results.clusterip }},docker-registry.default.svc,docker-registry.default.svc.cluster.local,{{ docker_registry_route_hostname }}"          --cert={{ openshift.common.config_base }}/master/registry.crt          --key={{ openshift.common.config_base }}/master/registry.key -        {% if openshift_version | oo_version_gte_3_5_or_1_5(openshift.common.deployment_type) | bool %}          --expire-days={{ openshift_hosted_registry_cert_expire_days | default(730) }} -        {% endif %}      - name: Update registry certificates secret        oc_secret: diff --git a/playbooks/common/openshift-cluster/sanity_checks.yml b/playbooks/common/openshift-cluster/sanity_checks.yml deleted file mode 100644 index 26716a92d..000000000 --- a/playbooks/common/openshift-cluster/sanity_checks.yml +++ /dev/null @@ -1,51 +0,0 @@ ---- -- name: Verify Requirements -  hosts: oo_all_hosts -  tasks: -  - fail: -      msg: Flannel can not be used with openshift sdn, set openshift_use_openshift_sdn=false if you want to use flannel -    when: openshift_use_openshift_sdn | default(true) | bool and openshift_use_flannel | default(false) | bool - -  - fail: -      msg: Nuage sdn can not be used with openshift sdn, set openshift_use_openshift_sdn=false if you want to use nuage -    when: openshift_use_openshift_sdn | default(true) | bool and openshift_use_nuage | default(false) | bool - -  - fail: -      msg: Nuage sdn can not be used with flannel -    when: openshift_use_flannel | default(false) | bool and openshift_use_nuage | default(false) | bool - -  - fail: -      msg: Contiv can not be used with openshift sdn, set openshift_use_openshift_sdn=false if you want to use contiv -    when: openshift_use_openshift_sdn | default(true) | bool and openshift_use_contiv | default(false) | bool - -  - fail: -      msg: Contiv can not be used with flannel -    when: openshift_use_flannel | default(false) | bool and openshift_use_contiv | default(false) | bool - -  - fail: -      msg: Contiv can not be used with nuage -    when: openshift_use_nuage | default(false) | bool and openshift_use_contiv | default(false) | bool - -  - fail: -      msg: Calico can not be used with openshift sdn, set openshift_use_openshift_sdn=false if you want to use Calico -    when: openshift_use_openshift_sdn | default(true) | bool and openshift_use_calico | default(false) | bool - -  - fail: -      msg: The Calico playbook does not yet integrate with the Flannel playbook in Openshift. Set either openshift_use_calico or openshift_use_flannel, but not both. -    when: openshift_use_calico | default(false) | bool and openshift_use_flannel | default(false) | bool - -  - fail: -      msg: Calico can not be used with Nuage in Openshift. Set either openshift_use_calico or openshift_use_nuage, but not both -    when: openshift_use_calico | default(false) | bool and openshift_use_nuage | default(false) | bool - -  - fail: -      msg: Calico can not be used with Contiv in Openshift. Set either openshift_use_calico or openshift_use_contiv, but not both -    when: openshift_use_calico | default(false) | bool and openshift_use_contiv | default(false) | bool - -  - fail: -      msg: openshift_hostname must be 63 characters or less -    when: openshift_hostname is defined and openshift_hostname | length > 63 - -  - fail: -      msg: openshift_public_hostname must be 63 characters or less -    when: openshift_public_hostname is defined and openshift_public_hostname | length > 63 diff --git a/playbooks/common/openshift-cluster/std_include.yml b/playbooks/common/openshift-cluster/std_include.yml deleted file mode 100644 index fe376fe31..000000000 --- a/playbooks/common/openshift-cluster/std_include.yml +++ /dev/null @@ -1,50 +0,0 @@ ---- -- name: Initialization Checkpoint Start -  hosts: all -  gather_facts: false -  roles: -  - installer_checkpoint -  tasks: -  - name: Set install initialization 'In Progress' -    run_once: true -    set_stats: -      data: -        installer_phase_initialize: -          status: "In Progress" -          start: "{{ lookup('pipe', 'date +%Y%m%d%H%M%SZ') }}" - -- include: evaluate_groups.yml -  tags: -  - always - -- include: initialize_facts.yml -  tags: -  - always - -- include: sanity_checks.yml -  tags: -  - always - -- include: validate_hostnames.yml -  tags: -  - node - -- include: initialize_openshift_repos.yml -  tags: -  - always - -- include: initialize_openshift_version.yml -  tags: -  - always - -- name: Initialization Checkpoint End -  hosts: all -  gather_facts: false -  tasks: -  - name: Set install initialization 'Complete' -    run_once: true -    set_stats: -      data: -        installer_phase_initialize: -          status: "Complete" -          end: "{{ lookup('pipe', 'date +%Y%m%d%H%M%SZ') }}" diff --git a/playbooks/common/openshift-cluster/upgrades/docker/docker_upgrade.yml b/playbooks/common/openshift-cluster/upgrades/docker/docker_upgrade.yml index 98953f72e..6d4ddf011 100644 --- a/playbooks/common/openshift-cluster/upgrades/docker/docker_upgrade.yml +++ b/playbooks/common/openshift-cluster/upgrades/docker/docker_upgrade.yml @@ -1,5 +1,5 @@  --- -- include: ../../evaluate_groups.yml +- include: ../../../../init/evaluate_groups.yml    vars:      # Do not allow adding hosts during upgrade.      g_new_master_hosts: [] diff --git a/playbooks/common/openshift-cluster/upgrades/init.yml b/playbooks/common/openshift-cluster/upgrades/init.yml index 6ad0b6b86..9981d905b 100644 --- a/playbooks/common/openshift-cluster/upgrades/init.yml +++ b/playbooks/common/openshift-cluster/upgrades/init.yml @@ -1,11 +1,11 @@  --- -- include: ../evaluate_groups.yml +- include: ../../../init/evaluate_groups.yml    vars:      # Do not allow adding hosts during upgrade.      g_new_master_hosts: []      g_new_node_hosts: [] -- include: ../initialize_facts.yml +- include: ../../../init/facts.yml  - name: Ensure firewall is not switched during upgrade    hosts: oo_all_hosts diff --git a/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml b/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml index 122066955..c634e0ab8 100644 --- a/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml +++ b/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml @@ -85,6 +85,8 @@    roles:    - openshift_manageiq +  - role: openshift_project_request_template +    when: openshift_project_request_template_manage    # Create the new templates shipped in 3.2, existing templates are left    # unmodified. This prevents the subsequent role definition for    # openshift_examples from failing when trying to replace templates that do @@ -117,7 +119,6 @@    - name: grep pluginOrderOverride      command: grep pluginOrderOverride {{ openshift.common.config_base }}/master/master-config.yaml      register: grep_plugin_order_override -    when: openshift.common.version_gte_3_3_or_1_3 | bool      changed_when: false      failed_when: false diff --git a/playbooks/common/openshift-cluster/upgrades/rpm_upgrade.yml b/playbooks/common/openshift-cluster/upgrades/rpm_upgrade.yml deleted file mode 100644 index 8cc46ab68..000000000 --- a/playbooks/common/openshift-cluster/upgrades/rpm_upgrade.yml +++ /dev/null @@ -1,38 +0,0 @@ ---- -# When we update package "a-${version}" and a requires b >= ${version} if we -# don't specify the version of b yum will choose the latest version of b -# available and the whole set of dependencies end up at the latest version. -# Since the package module, unlike the yum module, doesn't flatten a list -# of packages into one transaction we need to do that explicitly. The ansible -# core team tells us not to rely on yum module transaction flattening anyway. - -# TODO: If the sdn package isn't already installed this will install it, we -# should fix that -- name: Upgrade master packages -  package: name={{ master_pkgs | join(',') }} state=present -  vars: -    master_pkgs: -      - "{{ openshift.common.service_type }}{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-master{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-node{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-sdn-ovs{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-clients{{ openshift_pkg_version }}" -      - "tuned-profiles-{{ openshift.common.service_type }}-node{{ openshift_pkg_version }}" -      - PyYAML -  when: -    - component == "master" -    - not openshift.common.is_atomic | bool - -- name: Upgrade node packages -  package: name={{ node_pkgs | join(',') }} state=present -  vars: -    node_pkgs: -      - "{{ openshift.common.service_type }}{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-node{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-sdn-ovs{{ openshift_pkg_version }}" -      - "{{ openshift.common.service_type }}-clients{{ openshift_pkg_version }}" -      - "tuned-profiles-{{ openshift.common.service_type }}-node{{ openshift_pkg_version }}" -      - PyYAML -  when: -    - component == "node" -    - not openshift.common.is_atomic | bool diff --git a/playbooks/common/openshift-cluster/upgrades/upgrade_control_plane.yml b/playbooks/common/openshift-cluster/upgrades/upgrade_control_plane.yml index 8783ade99..fa65567c2 100644 --- a/playbooks/common/openshift-cluster/upgrades/upgrade_control_plane.yml +++ b/playbooks/common/openshift-cluster/upgrades/upgrade_control_plane.yml @@ -63,13 +63,9 @@    vars:      openshift_master_ha: "{{ groups.oo_masters_to_config | length > 1 }}"    serial: 1 -  handlers: -  - include: ../../../../roles/openshift_master/handlers/main.yml -    static: yes -  roles: -  - openshift_facts -  - lib_utils -  post_tasks: +  tasks: +  - include_role: +      name: openshift_facts    # Run the pre-upgrade hook if defined:    - debug: msg="Running master pre-upgrade hook {{ openshift_master_upgrade_pre_hook }}" @@ -78,55 +74,9 @@    - include: "{{ openshift_master_upgrade_pre_hook }}"      when: openshift_master_upgrade_pre_hook is defined -  - include: rpm_upgrade.yml component=master -    when: not openshift.common.is_containerized | bool - -  - include_vars: ../../../../roles/openshift_master_facts/vars/main.yml - -  - include: upgrade_scheduler.yml - -  - include: "{{ master_config_hook }}" -    when: master_config_hook is defined - -  - include_vars: ../../../../roles/openshift_master/vars/main.yml - -  - name: Update journald config -    include: ../../../../roles/openshift_master/tasks/journald.yml - -  - name: Remove any legacy systemd units and update systemd units -    include: ../../../../roles/openshift_master/tasks/systemd_units.yml - -  - name: Check for ca-bundle.crt -    stat: -      path: "{{ openshift.common.config_base }}/master/ca-bundle.crt" -    register: ca_bundle_stat -    failed_when: false - -  - name: Check for ca.crt -    stat: -      path: "{{ openshift.common.config_base }}/master/ca.crt" -    register: ca_crt_stat -    failed_when: false - -  - name: Migrate ca.crt to ca-bundle.crt -    command: mv ca.crt ca-bundle.crt -    args: -      chdir: "{{ openshift.common.config_base }}/master" -    when: ca_crt_stat.stat.isreg and not ca_bundle_stat.stat.exists - -  - name: Link ca.crt to ca-bundle.crt -    file: -      src: "{{ openshift.common.config_base }}/master/ca-bundle.crt" -      path: "{{ openshift.common.config_base }}/master/ca.crt" -      state: link -    when: ca_crt_stat.stat.isreg and not ca_bundle_stat.stat.exists - -  - name: Update oreg value -    yedit: -      src: "{{ openshift.common.config_base }}/master/master-config.yaml" -      key: 'imageConfig.format' -      value: "{{ oreg_url | default(oreg_url_master) }}" -    when: oreg_url is defined or oreg_url_master is defined +  - include_role: +      name: openshift_master +      tasks_from: upgrade.yml    # Run the upgrade hook prior to restarting services/system if defined:    - debug: msg="Running master upgrade hook {{ openshift_master_upgrade_hook }}" @@ -236,7 +186,6 @@      - reconcile_jenkins_role_binding_result.rc == 0      when:      - openshift_version | version_compare('3.7','<') -    - openshift_version | version_compare('3.4','>=')    - when: openshift_upgrade_target | version_compare('3.7','<')      block: @@ -368,13 +317,13 @@      delay: 60    roles: -  - lib_openshift    - openshift_facts -  - docker -  - openshift_node_dnsmasq -  - openshift_node_upgrade -    post_tasks: +  - include_role: +      name: openshift_node +      tasks_from: upgrade.yml +    vars: +      openshift_node_upgrade_in_progress: True    - name: Set node schedulability      oc_adm_manage_node:        node: "{{ openshift.node.nodename | lower }}" diff --git a/playbooks/common/openshift-cluster/upgrades/upgrade_nodes.yml b/playbooks/common/openshift-cluster/upgrades/upgrade_nodes.yml index c93a5d89c..5dc8193a7 100644 --- a/playbooks/common/openshift-cluster/upgrades/upgrade_nodes.yml +++ b/playbooks/common/openshift-cluster/upgrades/upgrade_nodes.yml @@ -34,16 +34,18 @@      delay: 60    roles: -  - lib_openshift    - openshift_facts -  - docker -  - openshift_node_dnsmasq -  - openshift_node_upgrade -  - role: openshift_excluder -    r_openshift_excluder_action: enable -    r_openshift_excluder_service_type: "{{ openshift.common.service_type }}" -    post_tasks: +  - include_role: +      name: openshift_node +      tasks_from: upgrade.yml +    vars: +      openshift_node_upgrade_in_progress: True +  - include_role: +      name: openshift_excluder +    vars: +      r_openshift_excluder_action: enable +      r_openshift_excluder_service_type: "{{ openshift.common.service_type }}"    - name: Set node schedulability      oc_adm_manage_node:        node: "{{ openshift.node.nodename | lower }}" diff --git a/playbooks/common/openshift-cluster/upgrades/upgrade_scheduler.yml b/playbooks/common/openshift-cluster/upgrades/upgrade_scheduler.yml deleted file mode 100644 index 8558bf3e9..000000000 --- a/playbooks/common/openshift-cluster/upgrades/upgrade_scheduler.yml +++ /dev/null @@ -1,173 +0,0 @@ ---- -# Upgrade predicates -- vars: -    prev_predicates: "{{ lookup('openshift_master_facts_default_predicates', short_version=openshift_upgrade_min, deployment_type=openshift_deployment_type) }}" -    prev_predicates_no_region: "{{ lookup('openshift_master_facts_default_predicates', short_version=openshift_upgrade_min, deployment_type=openshift_deployment_type, regions_enabled=False) }}" -    default_predicates_no_region: "{{ lookup('openshift_master_facts_default_predicates', regions_enabled=False) }}" -    # older_predicates are the set of predicates that have previously been -    # hard-coded into openshift_facts -    older_predicates: -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -      - name: NoVolumeZoneConflict -      - name: MaxEBSVolumeCount -      - name: MaxGCEPDVolumeCount -      - name: Region -        argument: -          serviceAffinity: -            labels: -            - region -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -      - name: NoVolumeZoneConflict -      - name: Region -        argument: -          serviceAffinity: -            labels: -            - region -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -      - name: Region -        argument: -          serviceAffinity: -            labels: -            - region -    # older_predicates_no_region are the set of predicates that have previously -    # been hard-coded into openshift_facts, with the Region predicate removed -    older_predicates_no_region: -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -      - name: NoVolumeZoneConflict -      - name: MaxEBSVolumeCount -      - name: MaxGCEPDVolumeCount -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -      - name: NoVolumeZoneConflict -    - - name: MatchNodeSelector -      - name: PodFitsResources -      - name: PodFitsPorts -      - name: NoDiskConflict -  block: - -  # Handle case where openshift_master_predicates is defined -  - block: -    - debug: -        msg: "WARNING: openshift_master_scheduler_predicates is set to defaults from an earlier release of OpenShift current defaults are: {{ openshift_master_scheduler_default_predicates }}" -      when: openshift_master_scheduler_predicates in older_predicates + older_predicates_no_region + [prev_predicates] + [prev_predicates_no_region] - -    - debug: -        msg: "WARNING: openshift_master_scheduler_predicates does not match current defaults of: {{ openshift_master_scheduler_default_predicates }}" -      when: openshift_master_scheduler_predicates != openshift_master_scheduler_default_predicates -    when: openshift_master_scheduler_predicates | default(none) is not none - -  # Handle cases where openshift_master_predicates is not defined -  - block: -    - debug: -        msg: "WARNING: existing scheduler config does not match previous known defaults automated upgrade of scheduler config is disabled.\nexisting scheduler predicates: {{ openshift_master_scheduler_current_predicates }}\ncurrent scheduler default predicates are: {{ openshift_master_scheduler_default_predicates }}" -      when: -      - openshift_master_scheduler_current_predicates != openshift_master_scheduler_default_predicates -      - openshift_master_scheduler_current_predicates not in older_predicates + [prev_predicates] - -    - set_fact: -        openshift_upgrade_scheduler_predicates: "{{ openshift_master_scheduler_default_predicates }}" -      when: -      - openshift_master_scheduler_current_predicates != openshift_master_scheduler_default_predicates -      - openshift_master_scheduler_current_predicates in older_predicates + [prev_predicates] - -    - set_fact: -        openshift_upgrade_scheduler_predicates: "{{ default_predicates_no_region }}" -      when: -      - openshift_master_scheduler_current_predicates != default_predicates_no_region -      - openshift_master_scheduler_current_predicates in older_predicates_no_region + [prev_predicates_no_region] - -    when: openshift_master_scheduler_predicates | default(none) is none - - -# Upgrade priorities -- vars: -    prev_priorities: "{{ lookup('openshift_master_facts_default_priorities', short_version=openshift_upgrade_min, deployment_type=openshift_deployment_type) }}" -    prev_priorities_no_zone: "{{ lookup('openshift_master_facts_default_priorities', short_version=openshift_upgrade_min, deployment_type=openshift_deployment_type, zones_enabled=False) }}" -    default_priorities_no_zone: "{{ lookup('openshift_master_facts_default_priorities', zones_enabled=False) }}" -    # older_priorities are the set of priorities that have previously been -    # hard-coded into openshift_facts -    older_priorities: -    - - name: LeastRequestedPriority -        weight: 1 -      - name: SelectorSpreadPriority -        weight: 1 -      - name: Zone -        weight: 2 -        argument: -          serviceAntiAffinity: -            label: zone -    # older_priorities_no_region are the set of priorities that have previously -    # been hard-coded into openshift_facts, with the Zone priority removed -    older_priorities_no_zone: -    - - name: LeastRequestedPriority -        weight: 1 -      - name: SelectorSpreadPriority -        weight: 1 -  block: - -  # Handle case where openshift_master_priorities is defined -  - block: -    - debug: -        msg: "WARNING: openshift_master_scheduler_priorities is set to defaults from an earlier release of OpenShift current defaults are: {{ openshift_master_scheduler_default_priorities }}" -      when: openshift_master_scheduler_priorities in older_priorities + older_priorities_no_zone + [prev_priorities] + [prev_priorities_no_zone] - -    - debug: -        msg: "WARNING: openshift_master_scheduler_priorities does not match current defaults of: {{ openshift_master_scheduler_default_priorities }}" -      when: openshift_master_scheduler_priorities != openshift_master_scheduler_default_priorities -    when: openshift_master_scheduler_priorities | default(none) is not none - -  # Handle cases where openshift_master_priorities is not defined -  - block: -    - debug: -        msg: "WARNING: existing scheduler config does not match previous known defaults automated upgrade of scheduler config is disabled.\nexisting scheduler priorities: {{ openshift_master_scheduler_current_priorities }}\ncurrent scheduler default priorities are: {{ openshift_master_scheduler_default_priorities }}" -      when: -      - openshift_master_scheduler_current_priorities != openshift_master_scheduler_default_priorities -      - openshift_master_scheduler_current_priorities not in older_priorities + [prev_priorities] - -    - set_fact: -        openshift_upgrade_scheduler_priorities: "{{ openshift_master_scheduler_default_priorities }}" -      when: -      - openshift_master_scheduler_current_priorities != openshift_master_scheduler_default_priorities -      - openshift_master_scheduler_current_priorities in older_priorities + [prev_priorities] - -    - set_fact: -        openshift_upgrade_scheduler_priorities: "{{ default_priorities_no_zone }}" -      when: -      - openshift_master_scheduler_current_priorities != default_priorities_no_zone -      - openshift_master_scheduler_current_priorities in older_priorities_no_zone + [prev_priorities_no_zone] - -    when: openshift_master_scheduler_priorities | default(none) is none - - -# Update scheduler -- vars: -    scheduler_config: -      kind: Policy -      apiVersion: v1 -      predicates: "{{ openshift_upgrade_scheduler_predicates -                      | default(openshift_master_scheduler_current_predicates) }}" -      priorities: "{{ openshift_upgrade_scheduler_priorities -                      | default(openshift_master_scheduler_current_priorities) }}" -  block: -  - name: Update scheduler config -    copy: -      content: "{{ scheduler_config | to_nice_json }}" -      dest: "{{ openshift_master_scheduler_conf }}" -      backup: true -  when: > -    openshift_upgrade_scheduler_predicates is defined or -    openshift_upgrade_scheduler_priorities is defined diff --git a/playbooks/common/openshift-cluster/upgrades/v3_6/master_config_upgrade.yml b/playbooks/common/openshift-cluster/upgrades/v3_6/master_config_upgrade.yml deleted file mode 100644 index db0c8f886..000000000 --- a/playbooks/common/openshift-cluster/upgrades/v3_6/master_config_upgrade.yml +++ /dev/null @@ -1,15 +0,0 @@ ---- -- modify_yaml: -    dest: "{{ openshift.common.config_base}}/master/master-config.yaml" -    yaml_key: 'controllerConfig.serviceServingCert.signer.certFile' -    yaml_value: service-signer.crt - -- modify_yaml: -    dest: "{{ openshift.common.config_base}}/master/master-config.yaml" -    yaml_key: 'controllerConfig.serviceServingCert.signer.keyFile' -    yaml_value: service-signer.key - -- modify_yaml: -    dest: "{{ openshift.common.config_base }}/master/master-config.yaml" -    yaml_key: servingInfo.clientCA -    yaml_value: ca.crt diff --git a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade.yml b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade.yml index dd109cfa9..ef52f214b 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade.yml @@ -63,7 +63,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_control_plane.yml b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_control_plane.yml index 8ab68002d..4c6646a38 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_control_plane.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_control_plane.yml @@ -67,7 +67,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_nodes.yml b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_nodes.yml index ba6fcc3f8..f25cfe0d0 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_nodes.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_6/upgrade_nodes.yml @@ -56,7 +56,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_7/master_config_upgrade.yml b/playbooks/common/openshift-cluster/upgrades/v3_7/master_config_upgrade.yml deleted file mode 100644 index 1d4d1919c..000000000 --- a/playbooks/common/openshift-cluster/upgrades/v3_7/master_config_upgrade.yml +++ /dev/null @@ -1,20 +0,0 @@ ---- -- modify_yaml: -    dest: "{{ openshift.common.config_base}}/master/master-config.yaml" -    yaml_key: 'controllerConfig.election.lockName' -    yaml_value: 'openshift-master-controllers' - -- modify_yaml: -    dest: "{{ openshift.common.config_base}}/master/master-config.yaml" -    yaml_key: 'controllerConfig.serviceServingCert.signer.certFile' -    yaml_value: service-signer.crt - -- modify_yaml: -    dest: "{{ openshift.common.config_base}}/master/master-config.yaml" -    yaml_key: 'controllerConfig.serviceServingCert.signer.keyFile' -    yaml_value: service-signer.key - -- modify_yaml: -    dest: "{{ openshift.common.config_base }}/master/master-config.yaml" -    yaml_key: servingInfo.clientCA -    yaml_value: ca.crt diff --git a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade.yml b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade.yml index f4862e321..e3c012380 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade.yml @@ -67,7 +67,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_control_plane.yml b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_control_plane.yml index b905d6d86..a88fa7b2e 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_control_plane.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_control_plane.yml @@ -71,7 +71,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_nodes.yml b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_nodes.yml index bc080f9a3..c0546bd2d 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_nodes.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_7/upgrade_nodes.yml @@ -56,7 +56,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade.yml b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade.yml index b3162bd5f..73df15d53 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade.yml @@ -67,7 +67,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_control_plane.yml b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_control_plane.yml index 3df5b17b5..48d55c16f 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_control_plane.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_control_plane.yml @@ -71,7 +71,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_nodes.yml b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_nodes.yml index f3d192ba7..abd56e762 100644 --- a/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_nodes.yml +++ b/playbooks/common/openshift-cluster/upgrades/v3_8/upgrade_nodes.yml @@ -56,7 +56,7 @@    tags:    - pre_upgrade -- include: ../../initialize_openshift_version.yml +- include: ../../../../init/version.yml    tags:    - pre_upgrade    vars: diff --git a/playbooks/common/openshift-cluster/validate_hostnames.yml b/playbooks/common/openshift-cluster/validate_hostnames.yml deleted file mode 100644 index be2e6a15a..000000000 --- a/playbooks/common/openshift-cluster/validate_hostnames.yml +++ /dev/null @@ -1,23 +0,0 @@ ---- -- name: Validate node hostnames -  hosts: oo_nodes_to_config -  tasks: -  - name: Query DNS for IP address of {{ openshift.common.hostname }} -    shell: -      getent ahostsv4 {{ openshift.common.hostname }} | head -n 1 | awk '{ print $1 }' -    register: lookupip -    changed_when: false -    failed_when: false -  - name: Warn user about bad openshift_hostname values -    pause: -      prompt: -        The hostname {{ openshift.common.hostname }} for {{ ansible_nodename }} -        doesn't resolve to an IP address owned by this host. Please set -        openshift_hostname variable to a hostname that when resolved on the host -        in question resolves to an IP address matching an interface on this -        host. This host will fail liveness checks for pods utilizing hostPorts, -        press ENTER to continue or CTRL-C to abort. -      seconds: "{{ 10 if openshift_override_hostname_check | default(false) | bool else omit }}" -    when: -    - lookupip.stdout != '127.0.0.1' -    - lookupip.stdout not in ansible_all_ipv4_addresses diff --git a/playbooks/common/openshift-master/additional_config.yml b/playbooks/common/openshift-master/additional_config.yml index 4fef5b923..32f638d42 100644 --- a/playbooks/common/openshift-master/additional_config.yml +++ b/playbooks/common/openshift-master/additional_config.yml @@ -21,6 +21,8 @@    roles:    - role: openshift_master_cluster      when: openshift_master_ha | bool and openshift.master.cluster_method == "pacemaker" +  - role: openshift_project_request_template +    when: openshift_project_request_template_manage    - role: openshift_examples      when: openshift_install_examples | default(true, true) | bool      registry_url: "{{ openshift.master.registry_url }}" diff --git a/playbooks/common/openshift-master/restart_services.yml b/playbooks/common/openshift-master/restart_services.yml index 4f8b758fd..4e1b3a3be 100644 --- a/playbooks/common/openshift-master/restart_services.yml +++ b/playbooks/common/openshift-master/restart_services.yml @@ -1,22 +1,4 @@  --- -- name: Restart master API -  service: -    name: "{{ openshift.common.service_type }}-master-api" -    state: restarted -  when: openshift_master_ha | bool -- name: Wait for master API to come back online -  wait_for: -    host: "{{ openshift.common.hostname }}" -    state: started -    delay: 10 -    port: "{{ openshift.master.api_port }}" -    timeout: 600 -  when: openshift_master_ha | bool -- name: Restart master controllers -  service: -    name: "{{ openshift.common.service_type }}-master-controllers" -    state: restarted -  # Ignore errrors since it is possible that type != simple for -  # pre-3.1.1 installations. -  ignore_errors: true -  when: openshift_master_ha | bool +- include_role: +    name: openshift_master +    tasks_from: restart.yml diff --git a/playbooks/common/openshift-master/scaleup.yml b/playbooks/common/openshift-master/scaleup.yml index 05b37d59f..4c415ebce 100644 --- a/playbooks/common/openshift-master/scaleup.yml +++ b/playbooks/common/openshift-master/scaleup.yml @@ -32,11 +32,7 @@    - name: verify api server      command: >        curl --silent --tlsv1.2 -      {% if openshift.common.version_gte_3_2_or_1_2 | bool %}        --cacert {{ openshift.common.config_base }}/master/ca-bundle.crt -      {% else %} -      --cacert {{ openshift.common.config_base }}/master/ca.crt -      {% endif %}        {{ openshift.master.api_url }}/healthz/ready      args:        # Disables the following warning: diff --git a/playbooks/common/openshift-master/tasks/wire_aggregator.yml b/playbooks/common/openshift-master/tasks/wire_aggregator.yml index 0d23e9d61..97acc5d5d 100644 --- a/playbooks/common/openshift-master/tasks/wire_aggregator.yml +++ b/playbooks/common/openshift-master/tasks/wire_aggregator.yml @@ -201,11 +201,7 @@    # wait_for port doesn't provide health information.    command: >      curl --silent --tlsv1.2 -    {% if openshift.common.version_gte_3_2_or_1_2 | bool %}      --cacert {{ openshift.common.config_base }}/master/ca-bundle.crt -    {% else %} -    --cacert {{ openshift.common.config_base }}/master/ca.crt -    {% endif %}      {{ openshift.master.api_url }}/healthz/ready    args:      # Disables the following warning: diff --git a/playbooks/common/openshift-node/image_prep.yml b/playbooks/common/openshift-node/image_prep.yml index 30651a1df..3c042acdc 100644 --- a/playbooks/common/openshift-node/image_prep.yml +++ b/playbooks/common/openshift-node/image_prep.yml @@ -1,15 +1,12 @@  ---  - name: normalize groups -  include: ../../byo/openshift-cluster/initialize_groups.yml - -- name: evaluate the groups -  include: ../openshift-cluster/evaluate_groups.yml +  include: ../../init/evaluate_groups.yml  - name: initialize the facts -  include: ../openshift-cluster/initialize_facts.yml +  include: ../../init/facts.yml  - name: initialize the repositories -  include: ../openshift-cluster/initialize_openshift_repos.yml +  include: ../../init/repos.yml  - name: run node config setup    include: setup.yml diff --git a/playbooks/common/openshift-node/network_manager.yml b/playbooks/common/openshift-node/network_manager.yml index b3a7399dc..c2efb0483 100644 --- a/playbooks/common/openshift-node/network_manager.yml +++ b/playbooks/common/openshift-node/network_manager.yml @@ -1,5 +1,5 @@  --- -- include: ../openshift-cluster/evaluate_groups.yml +- include: ../../init/evaluate_groups.yml  - name: Install and configure NetworkManager    hosts: oo_all_hosts | 
