Alerts reworked

Change alerts names, severities and descriptions.

Change-Id: Ib06f08a6f336d28592d5f70e97aedfeb12eb603c
Closes-bug: PROD-19698
This commit is contained in:
Michal Kobus 2018-05-03 15:28:45 +02:00
parent 38242186dd
commit 355aa0b480
2 changed files with 73 additions and 37 deletions

View file

@ -29,11 +29,8 @@
{%- set monitoring = salt['grains.filter_by']({ {%- set monitoring = salt['grains.filter_by']({
'default': { 'default': {
'remote_storage_adapter': { 'remote_storage_adapter': {
'sent_vs_received_ratio': 10.0, 'sent_vs_received_ratio': 0.9,
'ignored_vs_sent_ratio': 5.0, 'ignored_vs_sent_ratio': 0.05,
},
'alertmanager': {
'notification_failed_rate': 0.3
}, },
'prometheus': { 'prometheus': {
'remote_storage_queue_full_percent': 75.0, 'remote_storage_queue_full_percent': 75.0,
@ -51,3 +48,4 @@
'config_dir': '/srv/volumes/local/alerta', 'config_dir': '/srv/volumes/local/alerta',
}, },
}, merge=salt['pillar.get']('prometheus:alerta')) %}} }, merge=salt['pillar.get']('prometheus:alerta')) %}}

View file

@ -5,45 +5,55 @@ server:
{%- if server.get('enabled', False) %} {%- if server.get('enabled', False) %}
{% raw %} {% raw %}
PrometheusTargetDown: PrometheusTargetDown:
if: 'up != 1' if: up != 1
for: 2m for: 2m
labels: labels:
severity: critical severity: critical
service: prometheus service: prometheus
annotations: annotations:
summary: 'Prometheus endpoint {{ $labels.instance }} down' summary: "Prometheus target is down"
description: 'The Prometheus target {{ $labels.instance }} is down for the job {{ $labels.job }}.' description: "The Prometheus target for the {{ $labels.job }} job on the {{ $labels.host or $labels.instance }} node is down for at least 2 minutes."
PrometheusTargetSamplesOrderWarning:
if: increase(prometheus_target_scrapes_sample_out_of_order_total[1m]) > 0
labels:
severity: warning
service: prometheus
annotations:
summary: "Prometheus samples are out of order"
description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance are out of order."
PrometheusTargetSamplesBoundsWarning:
if: increase(prometheus_target_scrapes_sample_out_of_bounds_total[1m]) > 0
labels:
severity: warning
service: prometheus
annotations:
summary: "Prometheus samples timestamps are out of bounds"
description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance have timestamps out of bounds."
PrometheusTargetSamplesDuplicateWarning:
if: increase(prometheus_target_scrapes_sample_duplicate_timestamp_total[1m]) > 0
labels:
severity: warning
service: prometheus
annotations:
summary: "Prometheus samples have duplicate timestamps"
description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance have duplicate timestamps."
{% endraw %} {% endraw %}
{%- if server.version == 1.7 %} {%- if server.version == 1.7 %}
{% raw %} {% raw %}
PrometheusRushMode: PrometheusDataIngestionWarning:
if: 'prometheus_local_storage_rushed_mode != 0' if: prometheus_local_storage_rushed_mode != 0
for: 10m for: 10m
labels: labels:
severity: warning severity: warning
service: prometheus service: prometheus
annotations: annotations:
summary: 'Prometheus {{ $labels.instance }} in rush mode' summary: "Prometheus is in the rushed mode"
description: 'The Prometheus {{ $labels.instance }} is in rush mode for 10m.' description: "The Prometheus service writes on the {{ $labels.instance }} instance do not keep up with data ingestion speed for at least 10 minutes."
{% endraw %} {% endraw %}
{%- endif %} {%- endif %}
{%- endif %} {%- endif %}
{%- if alertmanager.get('enabled', False) %}
AlertmanagerNotificationFailed:
{%- set threshold = monitoring.alertmanager.notification_failed_rate|float %}
if: >-
rate(alertmanager_notifications_failed_total[5m]) > {{ threshold }}
for: 2m
{%- raw %}
labels:
severity: warning
service: alertmanager
annotations:
summary: 'Alertmanager {{ $labels.instance }} failed notifications'
description: 'Alertmanager {{ $labels.instance }} failed notifications for {{ $labels.integration }} (current value={{ $value }}, threshold={%- endraw %}{{ threshold }})'
{%- endif %}
{%- if server.get('config', {}).get('remote_write') %} {%- if server.get('config', {}).get('remote_write') %}
PrometheusRemoteStorageQueue: PrometheusRemoteStorageQueueFullWarning:
{%- set threshold = monitoring.prometheus.remote_storage_queue_full_percent|float %} {%- set threshold = monitoring.prometheus.remote_storage_queue_full_percent|float %}
if: >- if: >-
prometheus_remote_storage_queue_length / prometheus_remote_storage_queue_capacity * 100 > {{ threshold }} prometheus_remote_storage_queue_length / prometheus_remote_storage_queue_capacity * 100 > {{ threshold }}
@ -53,31 +63,59 @@ server:
severity: warning severity: warning
service: prometheus service: prometheus
annotations: annotations:
summary: 'Prometheus {{ $labels.instance }} remote storage queue is filling' summary: "Prometheus remote storage queue is full in {%- endraw %} {{ threshold }}{%- raw %}%"
description: 'The Prometheus {{ $labels.instance }} remote storage queue almost full (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%)' description: "The Prometheus remote storage queue on the {{ $labels.instance }} instance is {{ $value }}% full for at least 2 minutes."
{%- endraw %}
{%- endif %} {%- endif %}
{%- if remote_storage_adapter.get('enabled', False) %} {%- if remote_storage_adapter.get('enabled', False) %}
RemoteStorageAdapterSendingTooSlow: RemoteStorageAdapterMetricsSendingWarning:
{%- set threshold = monitoring.remote_storage_adapter.sent_vs_received_ratio|float %} {%- set threshold = monitoring.remote_storage_adapter.sent_vs_received_ratio|float %}
if: >- if: >-
100.0 - (100.0 * sent_samples_total{job="remote_storage_adapter"} / on (job, instance) received_samples_total) > {{ threshold }} increase(sent_samples_total{job="remote_storage_adapter"}[1m]) / on (job, instance) increase(received_samples_total[1m]) < {{ threshold }}
{% raw %} {% raw %}
labels: labels:
severity: warning severity: warning
service: remote_storage_adapter service: remote_storage_adapter
annotations: annotations:
summary: 'Remote storage adapter too slow on {{ $labels.instance }}' summary: "Ratio of sent to received remote storage adapter metrics is {%- endraw %} {{ threshold }}{%- raw %}"
description: 'Remote storage adapter can not ingest samples fast enough on {{ $labels.instance }} (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%).' description: "The ratio of the sent to received metrics of the remote storage adapter on the {{ $labels.instance }} instance is {{ $value }}."
RemoteStorageAdapterIgnoredTooHigh: {% endraw %}
RemoteStorageAdapterMetricsIgnoredWarning:
{%- set threshold = monitoring.remote_storage_adapter.ignored_vs_sent_ratio|float %} {%- set threshold = monitoring.remote_storage_adapter.ignored_vs_sent_ratio|float %}
if: >- if: >-
100.0 * prometheus_influxdb_ignored_samples_total{job="remote_storage_adapter"} / on (job, instance) sent_samples_total > {{ threshold }} increase(prometheus_influxdb_ignored_samples_total{job="remote_storage_adapter"}[1m]) / on (job, instance) increase(sent_samples_total[1m]) >= {{ threshold }}
{% raw %} {% raw %}
labels: labels:
severity: warning severity: warning
service: remote_storage_adapter service: remote_storage_adapter
annotations: annotations:
summary: 'Remote storage adapter receiving too many invalid metrics on {{ $labels.instance }}' summary: "{%- endraw %}{{ threshold }}{%- raw %}% of remote storage adapter metrics are invalid"
description: 'Remote storage adapter is receiving too many invalid metrics on {{ $labels.instance }} (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%).' description: "{{ $value }}% of remote storage adapter metrics on the {{ $labels.instance }} instance are invalid."
{%- endraw %}
{%- endif %}
{%- if alertmanager.get('enabled', False) %}
{%- raw %}
AlertmanagerNotificationFailureWarning:
if: >-
increase(alertmanager_notifications_failed_total[2m]) > 0
for: 2m
labels:
severity: warning
service: alertmanager
annotations:
summary: "Alertmanager notifications fail"
description: "An average of {{ $value }} Alertmanager {{ $labels.integration }} notifications on the {{ $labels.instance }} instance fail for at least 2 minutes."
AlertmanagerAlertsInvalidWarning:
if: >-
increase(alertmanager_alerts_invalid_total[2m]) > 0
for: 2m
labels:
severity: warning
service: alertmanager
annotations:
summary: "Alertmanager alerts are invalid"
description: "An average of {{ $value }} Alertmanager {{ $labels.integration }} alerts on the {{ $labels.instance }} instance are invalid for at least 2 minutes."
{%- endraw %}
{%- endif %} {%- endif %}
{%- endif %} {%- endif %}