| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|
|
|
| |
Lifted from https://github.com/openshift/origin/pull/8317
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
There are points where the docker service is restarted during
the install. Sometimes the services that are PartOf docker.service
do not get restarted when docker is restarted.
https://bugzilla.redhat.com/show_bug.cgi?id=1318948
Systemd documentation recommends using "wants" as the perferred
means of creating startup linkages between units. This
patch makes the ose services wanted by the docker service rather
than multi-user.target.
This creates a downward link from the docker service to the ose
containerized services.
|
|\
| |
| | |
Add /dev to node containers
|
| | |
|
|/
|
|
|
| |
Adds a new inventory var "openshift_node_local_quota_per_fsgroup", which will
set the perFSGroup quota node-config.yaml.
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This is the containerized openshift_pkg_version equivalent. Originally I was
hoping to reuse openshift_pkg_version for containerized installs but the fact
that it's very coupled to yum made that pretty ugly.
However, I did opt to rely on the previously existing 'openshift_version'
variable. Containerized and RPM installs can both use that variable and it
will be set appropriately if either openshift_pkg_version or
openshift_image_tag are set. I suspect someday containerized installs will be
the only option and I didn't can to have thinkgs like openshift_pkg_version and
openshift_image_tag in the playbooks anymore the necessary.
|
| |
|
| |
|
|
|
|
| |
Addressed review comments
|
|
|
|
|
|
| |
It's possible that openshift containers are started sooner
than openvswitch starts/sets networking (this typically
happens when docker service is restarted).
|
|\
| |
| | |
Nuage
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Added variables
Made changes for node configuration
Add service restart logic to node
Fixed ansible syntax errors
Add cert and key info for nodes
Added active and standby controller ip configuration information
Uncommented the nuage sdn check
Changed ca_crt -> ca_cert
Added restarting of atomic openshift master
Removed service account dependencies
Fixes
Fixed the api server url
Removed redundant restart of atomic openshift master
Configure nuagekubemon on all of the master nodes
Restart master api and controllers as well on nuagekubemon installation
Converted plugin config into template
Add template for nuagekubemon
Removed uplink interface from vars
Able to copy cert keys
Uninstall default ovs
Add the kubemon template
Do not install rdo sdn rpms in case of nuage
Addressed latest review comments
Set the networkPluginName for nuage
|
| | |
|
| | |
|
|/
|
|
| |
copied from https://github.com/eparis/kubernetes-ansible/blob/17f98edd7ff53e649b43e26822b8fbc0be42b233/roles/common/tasks/main.yml
|
| |
|
| |
|
| |
|
|
|
|
|
|
| |
- Attempt to detect the MTU of the interface associated with the node IP and
set the default for sdn_mtu accordingly. The value can still be overriden
by users and if detection fails the previous default of 1450 is still used.
|
|
|
|
|
|
| |
This is useful when we want to use a different networking solution
without having any networking plugin code run when the services start
up.
|
| |
|
|
|
|
|
|
| |
Previously this was being set to /var/lib/origin regardless of deployment_type
which isn't correct given that existing 'enterprise' and 'online' deployments
would have been deployed with /var/lib/openshift
|
|
|
|
|
| |
Sync's with Origin 1.0.6 and OSE 3.0.2 node-config.yaml networkConfig changes
Fixes #534
|
|\
| |
| | |
Cluster Metrics
|
| |
| |
| |
| |
| | |
Playbook based on
https://docs.openshift.org/latest/admin_guide/cluster_metrics.html.
|
| | |
|
| | |
|
|/
|
|
| |
apiServerArguments
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- Add support to bin/cluster for specifying etcd hosts
- defaults to 0, if no etcd hosts are selected, then configures embedded
etcd
- Updates for the byo inventory file for etcd and master as node by default
- Consolidation of cluster logic more centrally into common playbook
- Added etcd config support to playbooks
- Restructured byo playbooks to leverage the common openshift-cluster playbook
- Added support to common master playbook to generate and apply external etcd
client certs from the etcd ca
- start of refactor for better handling of master certs in a multi-master
environment.
- added the openshift_master_ca and openshift_master_certificates roles to
manage master certs instead of generating them in the openshift_master
role
- added etcd host groups to the cluster update playbooks
- aded better handling of host groups when they are either not present or are
empty.
- Update AWS readme
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Remove openshift-deployer.kubeconfig from master template
Sync config template
Update enterprise image names
Switch to node auto registration
Add deployer to list of serviceAccountConfig.managedNames
Move package installation before registering facts
change default kubeconfig location
Change system:openshift-client to system:openshift-master
Rename node cert/key/kubeconfig per openshift/origin#3160
Update references to /var/lib/openshift/openshift.local.certificates
|
|
- Templatize node config
- Templatize master config
- Integrated sdn changes
- Updates for openshift_facts
- Added support for node, master and sdn related changes
- registry_url
- added identity provider facts
- Removed openshift_sdn_* roles
- Install httpd-tools if configuring htpasswd auth
- Remove references to external_id
- Setting external_id interferes with nodes associating with the generated
node object when pre-registering nodes.
- osc/oc and osadm/oadm binary detection in openshift_facts
Misc Changes:
- make non-errata puddle default for byo example
- comment out master in list of nodes in inventory/byo/hosts
- remove non-error errors from fluentd_* roles
- Use admin kubeconfig instead of openshift-client
|