| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|\
| |
| | |
Bug 1527178 - installation of logging stack failed: Invalid version s…
|
| |
| |
| |
| |
| |
| |
| |
| | |
specified for Elasticsearch
openshift_logging_{curator,elasicsearch,fluentd,kibana,mux}/vars/main.yml:
- adding "3_9" to __allowed_.*_versions
- bumping __latest_.*_version to "3_9"
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Adding logic to do a full cluster restart if we are incrementing our …
…major versions of ES
This will help with the upgrade from 2.x to 5.x for ES, it also fixes something I came across with the handler on 3.7 where it checks the prior deployed version of the ES pod rather than the new one.
|
| |
| |
| |
| | |
full restart
|
| |
| |
| |
| | |
versions of ES
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
It appears that when one role dynamically imports
another, usage of import_role inside the dynamically
included role is not possible.
If something is included with include_role (dynamic),
all tasks therein must also use include_role (dynamic).
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
In Ansible 2.2, the include_role directive came into existence as
a Tech Preview. It is still a Tech Preview through Ansible 2.4
(and in current devel branch), but with a noteable change. The
default behavior switched from static: true to static: false
because that functionality moved to the newly introduced
import_role directive (in order to stay consistent with include*
being dynamic in nature and `import* being static in nature).
The dynamic include is considerably more memory intensive as it will
dynamically create a role import for every host in the inventory
list to be used. (Also worth noting, there is at the time of this
writing an object allocation inefficiency in the dynamic include
that can in certain situations amplify this effect considerably)
This change is meant to mitigate the pressure on memory for the
Ansible control host.
We need to evaluate where it makes sense to dynamically include roles
and revert back to dynamic inclusion if and where it makes sense to do
so.
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Update logging to use existing cluster deployment for defaults
This will allow us to use logging facts to set defaults of specific configurations such as ES index replicas and shard count.
The update to logging facts yields us output like:
```json
"elasticsearch": {
"clusterrolebindings": {},
"configmaps": {
"logging-elasticsearch": {
"elasticsearch.yml": {
"cloud": {
"kubernetes": {
"namespace": "${NAMESPACE}",
"pod_label": "${POD_LABEL}",
"pod_port": 9300
}
},
"cluster": {
"name": "${CLUSTER_NAME}"
},
"discovery": {
"type": "kubernetes",
"zen.minimum_master_nodes": "${NODE_QUORUM}",
"zen.ping.multicast.enabled": false
},
"gateway": {
"expected_nodes": "${RECOVER_EXPECTED_NODES}",
"recover_after_nodes": "${NODE_QUORUM}",
"recover_after_time": "${RECOVER_AFTER_TIME}"
},
"index": {
"number_of_replicas": 0,
"number_of_shards": 1,
"translog": {
"flush_threshold_period": "5m",
"flush_threshold_size": "256mb"
},
"unassigned.node_left.delayed_timeout": "2m"
},
"io.fabric8.elasticsearch.authentication.users": [
"system.logging.kibana",
"system.logging.fluentd",
"system.logging.curator",
"system.admin"
],
```
TODO:
- [x] Update logging facts to pull out settings from config maps
- [x] Move `openshift_sanitize_inventory/library/conditional_set_fact.py` up to repo level
- [x] Generate diffs against currently deployed configs and correctly patch in custom changes from customers
- [x] Use `conditional_set_fact` to easily set defaults for logging based on logging facts, or falling back to role defaults when not specified in the inventory
- [x] Update all components to follow patching configmaps
|
| |
| |
| |
| | |
files, use diffs to keep custom changes, white list certain settings when creating diffs
|
|/
|
|
|
|
|
|
| |
specified for Elasticsearch
openshift_logging_{curator,elasicsearch,fluentd,kibana,mux}/vars/main.yml:
- adding "3_8" to __allowed_.*_versions
- replacing the value of __latest_.*_version "3_6" with "3_8".
|
|
|
|
|
|
|
| |
Move openshift_deployment_type check into sanity_check
action plugin. Remove compatibility for deployment_type.
deployment_type has been deprecated for some time now.
|
|
|
|
|
| |
We set these variables using facts in init, no need
to duplicate the logic all around the codebase.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
This commit relocates filter_plugings to lib_utils,
changes the namespacing to prevent unintended use of
older versions that may be present in filter_plugins/
directory on existing installs.
Add lib_utils to meta depends for roles
Also consolidate some plugins into lib_utils from
various other areas.
Update rpm spec, obsolete plugin rpms.
|
|
|
|
| |
at the end of a deployment, allowing for override with variable
|
|\
| |
| | |
bug 1519622. Disable rollback of ES DCs
|
| | |
|
| | |
|
|/
|
|
|
|
| |
Remove hosted vars from openshift_facts.
The current pattern is causing a bunch of undesired sideffects.
|
| |
|
| |
|
|\
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
logging with static pvc: allow specifying the storage class name
See https://github.com/openshift/openshift-ansible/issues/6028 for more info and motivation.
|
| | |
|
| | |
|
| | |
|
|/
|
|
| |
openshift_logging pattern
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
bug 1506073. Lower cpu request for logging when it exceeds limit
This PR fixes https://bugzilla.redhat.com/show_bug.cgi?id=1506073 by:
* Lowering the CPU request to match the limit when the request is greater then a specified limit
I have an open question on if this is an acceptable change of if it makes the outcome unexpected. Should we prefer to exit during the deployment and advise the operator to correct their inventory?
|
| | |
|
|/ |
|
|\
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Bug 1452939 - change imagePullPolicy in logging and metrics
cc: @jcantrill
|
| |
| |
| |
| |
| | |
- all images logging and metrics change their default imagePullPolicy
from Always to IfNotPresent
|
|\ \
| |/
|/| |
Updating to use same image as origin until enterprise image is built
|
| |
| |
| |
| | |
specified
|
|\ \
| | |
| | | |
Use "requests" for CPU resources instead of limits
|
| |/
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
We now use a CPU request to ensure logging infrastructure pods are
not capped by default for CPU usage. It is still important to ensure
we have a minimum amount of CPU.
We keep the use of the variables *_cpu_limit so that the existing
behavior is maintained.
Note that we don't want to cap an infra pod's CPU usage by default,
since we want to be able to use the necessary resources to complete
it's tasks.
Bug 1501960 (https://bugzilla.redhat.com/show_bug.cgi?id=1501960)
|
|\ \
| |/
|/| |
bug 1489498. preserve replica and shard settings
|
| | |
|
| | |
|
| | |
|
| | |
|
|/ |
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Bug 1496271 - Perserve SCC for ES local persistent storage
ES can be modified to use node local persistent storage. This requires changing SCC and is described in docs:
https://docs.openshift.com/container-platform/3.6/install_config/aggregate_logging.html
During an upgrade, SCC defined by the user is ignored. This fix fetches SCC user defined as a fact and adds it to the ES DC which is later used.
Also includes cherrypicked fix for - Bug 1482661 - Preserve ES dc nodeSelector and supplementalGroups
cc @jcantrill
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
ES can be modified to use node local persistent storage. This requires
changing SCC and is described in docs:
https://docs.openshift.com/container-platform/3.6/install_config/aggregate_logging.html
During an upgrade, SCC defined by the user is ignored. This fix fetches
SCC user defined as a fact and adds it to the ES DC which is later used.
|
| |
| |
| |
| | |
(cherry picked from commit 601e35cbf4410972c7fa0a1d3d5c6327b82353ac)
|
|\ \
| | |
| | |
| | |
| | |
| | |
| | | |
Automatic merge from submit-queue.
Add logging es prometheus endpoint
This PR adds changes to add a prometheus endpoint to the logging elasticsearch pod
|
| |/ |
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
PR https://github.com/openshift/openshift-ansible/pull/3509 has removed any
usage of `openshift_logging_es_cpu_limit`.
Currently, the `openshift_logging_elasticsearch_cpu_limit` is either default
'1000m' or derived from `openshift_logging_es_ops_cpu_limit` but if user sets
the `openshift_logging_es_cpu_limit` in the inventory as documented, its value
is ignored.
This PR fixes the issue by trying to set
openshift_logging_elasticsearch_cpu_limit=openshift_logging_es_cpu_limit
And including the role as -ops overrides this setting.
|
|\
| |
| | |
Merged by openshift-bot
|
| | |
|
|\ \
| | |
| | | |
Merged by openshift-bot
|