|
6 éve | |
---|---|---|
.. | ||
defaults | 6 éve | |
files | 7 éve | |
meta | 6 éve | |
tasks | 6 éve | |
templates | 6 éve | |
tests | 7 éve | |
vars | 6 éve | |
OWNERS | 7 éve | |
README.md | 6 éve |
OpenShift Prometheus Installation
For default values, see defaults/main.yaml
.
openshift_prometheus_state
: present - install/update. absent - uninstall.
openshift_prometheus_node_exporter_install
: true (default) or false
openshift_prometheus_namespace
: project (i.e. namespace) where the components will be
deployed.
openshift_prometheus_node_selector
: Selector for the nodes prometheus will be deployed on.
openshift_prometheus_args
: Modify or add arguments for prometheus application
openshift_prometheus_hostname
: specify the hostname for the route to prometheus prometheus-{{openshift_prometheus_namespace}}.{{openshift_master_default_subdomain}}
openshift_prometheus_alerts_hostname
: specify the hostname for the route to prometheus-alerts prometheus_alerts-{{openshift_prometheus_namespace}}.{{openshift_master_default_subdomain}}
e.g
openshift_prometheus_args=['--storage.tsdb.retention=6h', '--query.timeout=2m']
Each prometheus component (prometheus, alertmanager, alertbuffer) can set pv claim by setting corresponding role variable:
openshift_prometheus_<COMPONENT>_storage_type: <VALUE> (pvc, emptydir)
openshift_prometheus_<COMPONENT>_storage_class: <VALUE>
openshift_prometheus_<COMPONENT>_pvc_(name|size|access_modes|pv_selector): <VALUE>
e.g
openshift_prometheus_storage_type: pvc
openshift_prometheus_storage_class: glusterfs-storage
openshift_prometheus_alertmanager_pvc_name: alertmanager
openshift_prometheus_alertbuffer_pvc_size: 10G
openshift_prometheus_pvc_access_modes: [ReadWriteOnce]
NOTE: Setting openshift_prometheus_<COMPONENT>_storage_labels
overrides openshift_prometheus_<COMPONENT>_pvc_pv_selector
An external file with alert rules can be added by setting path to additional rules variable:
openshift_prometheus_additional_rules_file: <PATH>
File content should be in prometheus alert rules format. Following example sets rule to fire an alert when one of the cluster nodes is down:
groups:
- name: example-rules
interval: 30s # defaults to global interval
rules:
- alert: Node Down
expr: up{job="kubernetes-nodes"} == 0
annotations:
miqTarget: "ContainerNode"
severity: "HIGH"
message: "{{ '{{' }}{{ '$labels.instance' }}{{ '}}' }} is down"
Each prometheus component (prometheus, alertmanager, alert-buffer, oauth-proxy) can specify a cpu and memory limits and requests by setting the corresponding role variable:
openshift_prometheus_<COMPONENT>_(memory|cpu)_(limit|requests): <VALUE>
e.g
openshift_prometheus_alertmanager_memory_limit: 1Gi
openshift_prometheus_oath_proxy_cpu_request: 100
openshift_prometheus_node_exporter_cpu_limit: 200m
openshift_facts
- name: Configure openshift-prometheus
hosts: oo_first_master
roles:
- role: openshift_prometheus
Apache License, Version 2.0