| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|
|
|
|
|
|
|
|
| |
openshift.common.version
This would fix errors like 'Master running 3.9.0 must be upgraded to
3.9* before node upgrade can be run' during the upgrade
Signed-off-by: Vadim Rutkovsky <vrutkovs@redhat.com>
|
|\
| |
| |
| |
| | |
Automatic merge from submit-queue.
Bug 1536217: Need to validate etcd serving certs before 3.9 upgrade
|
| |
| |
| |
| | |
etcd serving cert SAN.
|
|/
|
|
|
|
|
|
|
|
| |
Master version is unprotected during all upgrades.
During node-only upgrades, master version should
remain protected to ensure correct image tag is
set for node upgrades.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1536839
|
|
|
|
|
|
|
|
| |
There are some obsoletes in 3.9 packages which lead yum to upgrade
directly from 3.7 to 3.9 when you ask for 3.8 packages while 3.9 repos
are enabled. Since we'd like to allow people to run one playbook to
upgrade from 3.7 to 3.8 to 3.9 we need to exclude those packages when
upgrading to 3.8
|
|
|
|
|
| |
This commit limits common init code to exclude
oo_nodes_to_config during upgrade_control_plane runs.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
In Ansible 2.2, the include_role directive came into existence as
a Tech Preview. It is still a Tech Preview through Ansible 2.4
(and in current devel branch), but with a noteable change. The
default behavior switched from static: true to static: false
because that functionality moved to the newly introduced
import_role directive (in order to stay consistent with include*
being dynamic in nature and `import* being static in nature).
The dynamic include is considerably more memory intensive as it will
dynamically create a role import for every host in the inventory
list to be used. (Also worth noting, there is at the time of this
writing an object allocation inefficiency in the dynamic include
that can in certain situations amplify this effect considerably)
This change is meant to mitigate the pressure on memory for the
Ansible control host.
We need to evaluate where it makes sense to dynamically include roles
and revert back to dynamic inclusion if and where it makes sense to do
so.
|
|
|
|
|
|
|
| |
Move openshift_deployment_type check into sanity_check
action plugin. Remove compatibility for deployment_type.
deployment_type has been deprecated for some time now.
|
|
|
|
|
| |
We set these variables using facts in init, no need
to duplicate the logic all around the codebase.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This commit relocates filter_plugings to lib_utils,
changes the namespacing to prevent unintended use of
older versions that may be present in filter_plugins/
directory on existing installs.
Add lib_utils to meta depends for roles
Also consolidate some plugins into lib_utils from
various other areas.
Update rpm spec, obsolete plugin rpms.
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This commit refactors some upgrade code paths.
Touched areas are:
1) Reduces usage of 'oo_all_hosts' in various places,
especially when running upgrade_control_plane.
2) Reuses common code across the various upgrade*
playbooks.
3) Moves docker upgrade checks into container_runtime_role.
4) Combines smaller playbooks and plays to reduce file sprawl.
|
|\
| |
| | |
Remove openshift.common.service_type
|
| |
| |
| |
| |
| |
| |
| |
| | |
This commit removes openshift.common.service_type
in favor of openshift_service_type.
This commit also removes r_openshift_excluder_service_type
from plays in favor of using the role's defaults.
|
|/ |
|
| |
|
|
|
|
|
| |
This commit removes openshift.common.cli_image in
favor of openshift_cli_image.
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Enable oreg_auth credential replace during upgrades
Currently, upgrades run a docker image pull prior to
upgrading masters and nodes for containerized installs.
If using a secure registry, and a user wishes to upgrade
their credentials due to expiry, the image pull will fail.
This commit ensures docker login credentials are updated
during upgrades, if necessary.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1503995
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Currently, upgrades run a docker image pull prior to
upgrading masters and nodes for containerized installs.
If using a secure registry, and a user wishes to upgrade
their credentials due to expiry, the image pull will fail.
This commit ensures docker login credentials are updated
during upgrades, if necessary.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1503995
|
|\ \
| | |
| | |
| | |
| | | |
ingvagabund/strip-dash-when-comparing-versions-python3
strip dash when comparing version with Python3
|
| |/ |
|
|/ |
|
|
|
|
| |
https://bugzilla.redhat.com/show_bug.cgi?id=1484324
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue
Default openshift_pkg_version to full version-release during upgrades
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1490677
The versioning scheme for 3.7 pre-releases has changed and now all
versions are 3.7.0 and the release is incremented on builds, ie:
3.7.0-0.124.0 upgraded to 3.7.0-0.125.0. If we know we're an upgrade and
they haven't requested a specific package version defer the defaulting
of openshift_pkg_version until the upgrade playbooks and there set it to
the available version including the release.
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1490677
The versioning scheme for 3.7 pre-releases has changed and now all
versions are 3.7.0 and the release is incremented on builds, ie:
3.7.0-0.124.0 upgraded to 3.7.0-0.125.0. If we know we're an upgrade and
they haven't requested a specific package version defer the defaulting
of openshift_pkg_version until the upgrade playbooks and there set it to
the available version including the release.
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
juanvallejo/jvallejo/add-additonal-checks-upgrade-path
Automatic merge from submit-queue
Adding additonal checks upgrade path
Depends on https://github.com/openshift/openshift-ansible/pull/4960
TODO
- Possibly handle `upgrade` playbook context on `etcd_volume` check
cc @sosiouxme @rhcarvalho
|
| | |
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Previously, openshift-ansible supported various
types of deployments using the variable "openshift_deployment_type"
Currently, openshift-ansible only supports two deployment types,
"origin" and "openshift-enterprise".
This commit removes all logic and references to deprecated
deployment types.
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Moved the checks for osm_cluster_network_cidr, osm_host_subnet_length,
openshift_portal_net from upgrade to openshift_sanitize_inventory
as we now consider it a required variable for install, updrade, or
scale up.
Signed-off-by: Steve Milner <smilner@redhat.com>
|
| |
| |
| |
| | |
Signed-off-by: Steve Milner <smilner@redhat.com>
|
|\ \
| | |
| | | |
upgrade: Verify required network items are set
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
When upgrading osm_cluster_network_cidr, osm_host_subnet_length, and
openshift_portal_net must be set to avoid SDN initialization errors.
This was found when the default parameters were changed between Openshift
versions. This meant users who upgraded and did not specify either
mentioned variable at install/upgrade time ended up getting SDN errors
post upgrade.
When osm_cluster_network_cidr, osm_host_subnet_length, and
openshift_portal_net are not set the upgrade will fail telling the user
that the variables must be set and how to find the current values in the
current install.
References: https://github.com/openshift/openshift-ansible/commit/b50b4ea0b03feb9431abd7294fe4fb6b549ddfc0
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1451023
|
|/ / |
|
| | |
|
|/ |
|
|
|
|
|
|
| |
From now on, all master configurations use the api / controller split,
regardless of HA mode or previous configuration. This will be our only
supported configuration starting in 3.7 or 3.8.
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
| |
- check both available excluder versions are at most of upgrade target version
- get excluder status through status command
- make excluders enablement configurable
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Building off the work done for Docker 1.10, we now require Docker 1.12
by default.
The upgrade process was already set to ensure you are running the latest
docker during upgrade, and the standalone docker upgrade playbook can
also be used if desired.
As before, you can override this Docker 1.12 requirement by setting a
docker_version=1.10.3 (or similar), and you can skip the default to
upgrade docker by setting docker_upgrade=False.
|