| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
| |
There is currently a bug in registry auth
credential creation logic for openshift_node
and openshift_node_upgrade
This commit fixes the logic.
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue
Move additional/block/insecure registires to /etc/containers/...
Move additional/block/insecure registires to /etc/containers/registries.conf
This commit moves additional/block/insecure registries to
/etc/containers/registries.conf and comments existing lines in
/etc/sysconfig/docker.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1460930
|
| |
| |
| |
| |
| |
| |
| |
| | |
This commit moves additional/block/insecure registries to
/etc/containers/registries.conf and comments existing lines in
/etc/sysconfig/docker.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1460930
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Currently, the logic for registry authentication is
not implemented correctly to account for upgrades of
containerized hosts.
Additionally, the logic to account for multiple runs
of openshift-ansible might cause registry authentication
credentials to not be mounted inside of containerized hosts.
This commit adds the necessary logic to ensure containerized
hosts retain registry credentials.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1494470
|
|
|
|
|
|
|
|
|
| |
Most of this role's purpose was to set facts. The vast majority
of these facts were simply redefining user-supplied variables.
This commit also removes various artifacts leftover from
previous versions, as well as variables that seem to be
entirely unused.
|
|
|
|
| |
Prevents playbooks from accidentally restarting the master service.
|
| |
|
|
|
|
| |
openshift_node_upgrade
|
|
|
|
|
|
|
|
|
|
|
| |
Most occurrences are in a form:
```yaml
openshift.common.use_openshift_sdn | default(true) | bool
```
Let's make all occurences this way given the use_openshift_sdn is set to true anyway.
See https://github.com/openshift/openshift-ansible/blob/0c350dcc7d06d62be5ba3a8e468dff85cdd96dd7/roles/openshift_facts/library/openshift_facts.py#L2035
|
|
|
|
|
|
| |
On nodes with thousands of services it may take a very long time to
establish all of the network routing rules. The longest we've seen is
about 180s
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|\
| |
| |
| |
| | |
ingvagabund/increase-timeout-when-waiting-for-master-and-nodes
increase implicit 300s default timeout to explicit 600s
|
| | |
|
| | |
|
| | |
|
| | |
|
|/ |
|
|\
| |
| | |
Retry docker startup once
|
| |
| |
| |
| |
| | |
Large environments frequently see docker fail to start up after an
upgrade. Not sure why but we can retry once.
|
|/ |
|
|
|
|
|
|
|
|
| |
When rpm_upgrade is used for upgrading the node's unit file will be
overridden. systemd is then reloaded if the files are templated
successfully.
Ref: https://bugzilla.redhat.com/show_bug.cgi?id=1455843
|
|
|
|
|
|
|
|
|
| |
The installer now lays down updated systemd unit files in
/etc/systemd/system/ which override the units that come with the
rpms. Because of this we need to reload systemd before restarting
services.
Ref: https://bugzilla.redhat.com/show_bug.cgi?id=1455843
|
|
|
|
|
|
|
|
|
| |
This commit allows to specify imageConfig.format specifically for master
or for nodes.
One use case of this could be if you want to use customer builder
images. In this case imageConfig.format only needs to be changed in the
master-config.yml but not in the node-config.yml.
|
|\
| |
| | |
Ensure good return code for specific until loops
|
| | |
|
|\ \
| | |
| | | |
Merged by openshift-bot
|
| | |
| | |
| | |
| | | |
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1450167
|
|/ /
| |
| |
| |
| |
| |
| |
| |
| | |
Sometimes the node container is not started on a container-engine
restart. Use a weaker dependency on openvswitch that is causing this issue
Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1451192
Signed-off-by: Giuseppe Scrivano <gscrivan@redhat.com>
|
|/
|
|
|
|
|
| |
- If using a system container: container-engine
- If using a package install: docker
Ref: https://bugzilla.redhat.com/show_bug.cgi?id=1448800
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This change allows for the use of either the traditional package install
of docker OR a system container install of docker.
Two new inventory options, openshift_docker_use_system_container and
openshift_docker_systemcontainer_image_registry_override, have been
added which are be used to install with a system container. By default
this option is commented out.
One new fact has been added:
- docker.service_name: docker by default, container-engine-docker
for system container
|
| |
|
|
|
|
|
|
| |
* Lines are commented out vice being removed
* Comment is added indicating why the change happened
* Variable openshift_disable_swap added to allow user control
|
|\
| |
| | |
Stop all services prior to upgrading, start all services after
|
| | |
|
|/ |
|
| |
|
|
|
|
|
|
| |
It appears that openshift was holding a lock or perhaps openvswitch on
its own does not restart cleanly when upgrading the package so just
stop it before updating the package.
|
|\
| |
| | |
Merged by openshift-bot
|
| | |
|
|\ \
| | |
| | | |
Merged by openshift-bot
|
| |/
| |
| |
| |
| |
| | |
Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1420636
Signed-off-by: Giuseppe Scrivano <gscrivan@redhat.com>
|
|/
|
|
| |
Fixes https://bugzilla.redhat.com/show_bug.cgi?id=1427807
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
around across node restart
With the move to a CNI plugin, docker no longer handles IPAM, but CNI does through
openshift-sdn's usage of the 'host-local' CNI IPAM plugin. That plugin stores
IPAM allocations under /var/lib/cni/.
If the node container gets restarted, without presreving /var/lib/cni, the IPs
currently allocated to running pods get lost and on restart, openshift-sdn
may allocate those IPs to new pods causing duplicate allocations.
This never happened with docker because it has its own persistent IPAM store that
does not get removed when docker restarts. Also because (historically) when docker
restarted, all the containers died and the IP allocations were released by the
daemon.
Fix this by ensuring that IPAM allocations (which are tied to the life of the pod,
*not* the life of the openshift-node process) persist even if the openshift-node
process restarts.
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1427789
|