Skip to content

Latest commit

 

History

History
322 lines (229 loc) · 14.1 KB

alerts.md

File metadata and controls

322 lines (229 loc) · 14.1 KB

Alerts

Summary

Kubecost alerts allow teams to receive updates on real-time Kubernetes spend. They are configurable via the Kubecost UI or Helm values. This resource gives an overview of how to configure Kubecost email and Slack alerts using Kubecost helm chart values.

As of v1.72.0, Kubecost supports four types of notifications:

  1. Recurring update - sends an email and/or Slack alert with cluster spend across all or a set of namespaces, with costs broken down by namespace

  2. Budget -- sends an email and/or Slack alert when spend crosses a defined threshold

  3. Spend Change -- sends an email and/or Slack alert reporting unexpected spend increases relative to moving averages

  4. [Beta] Efficiency -- detect when a Kubernetes tenant is operating below a target cost-efficiency threshold

  5. Kubecost Health Diagnostic -- used for production monitoring for the health of Kubecost itself

  6. Cluster Health -- used to determine if the cluster's health score changes by a specific threshold.

Have questions or issues? View our troubleshooting guide.

Configuring Alerts in Helm

Note: values.yaml is a source of truth. Alerts set through values.yaml will continually overwrite any manual alert settings set through the Kubecost UI.

Global Alert Parameters

The alert settings, under global.notifications.alertConfigs in cost-analyzer/values.yaml, accept four global fields:

  • enabled determines whether Kubecost will schedule and display the configured alerts in Notifications, default set to false
  • frontendUrl optional, your cost analyzer front end URL used for linkbacks in alert bodies
  • globalSlackWebhookUrl optional, a global Slack webhook used for alerts, enabled by default if provided
  • globalAlertEmails a global list of emails for alerts

Example Helm values.yaml:

notifications:
	alertConfigs:
		frontendUrl: http://localhost:9090  # optional, used for linkbacks
		globalSlackWebhookUrl: https://hooks.slack.com/services/T00000000/B00000000/XXXXXXXXXXXXXXXXXXXXXXXX  # optional, used for Slack alerts
		globalAlertEmails:
			- [email protected]
			- [email protected]
		alerts:
		# list of individual alerts
		...

Configuring Each Alert Type

In addition to globalSlackWebhookUrl and globalAlertEmails fields, every alert allows optional individual ownerContact (a list of email addresses) and slackWebhookUrl (if different from globalSlackWebhookUrl) fields. Alerts will default to the global Slack and email settings if these optional fields are not supplied.

Type: Recurring Update

Sends a recurring email and/or Slack alert with a summary report of cost and efficiency metrics.

Required parameters:

  • type: recurringUpdate
  • aggregation: <aggregation> -- configurable, accepts a single valid aggregation parameter*
  • filter: '*'
  • window: <N>d -- configurable, N ≥ 1

Valid Aggregation Parameters:

  • cluster
  • container
  • controller
  • namespace
  • pod
  • service
  • deployment
  • daemonset
  • statefulset
  • job
  • label requires the following format: label:<label_name>
  • annotation requires the following format: annotation:<annotation_name>

Required parameters (by individual namespace):

  • type: recurringUpdate
  • aggregation: namespace
  • filter: <value> -- configurable, accepts a single namespace name (comma-separated values unsupported)
  • window: 7d

Example Helm values.yaml:

			# Recurring weekly namespace update on all namespaces
			- type: recurringUpdate
				window: weekly  # or 7d
				aggregation: namespace
				filter: '*'
			# Recurring weekly namespace update on kubecost namespace
			- type: recurringUpdate
				window: weekly  # or 7d
				aggregation: namespace
				filter: kubecost
				ownerContact: # optional, overrides globalAlertEmails default
					- [email protected]
					- [email protected]
				slackWebhookUrl: https://hooks.slack.com/services/<different-from-global> # optional, overrides globalSlackWebhookUrl default

Type: Efficiency

Note: this feature is currently in Beta

Alert when Kubernetes tenants, e.g. namespaces or label sets, are running below defined cost-efficiency thresholds.

Required parameters:

  • type: efficiency
  • efficiencyThreshold: <threshold> -- efficiency threshold ranging from 0.0 to 1.0
  • aggregation: <agg-parameter> -- configurable, accepts all aggregations supported by the aggregated cost model API
  • window: <N>d number of days for measuring efficiency

Optional parameters:

  • filter: <value> -- limit the aggregations that this alert will cover, accepts comma-separated values
  • spendThreshold represents a minimum spend threshold for alerting

The example below sends a Slack alert when any namespace spending is running below 40% cost efficiency and has spent more than $100 during the last day.

- type: efficiency
	efficiencyThreshold: 0.4  # Alert if below this percentage cost efficiency
	spendThreshold: 100 # optional, alert if tenant has spend more than $100 over this window
	window: 1d    # measure efficiency over last 
	aggregation: namespace
	slackWebhookUrl: ‘https://hooks.slack.com/services/TE6GRBNET/BFFK0P848/jFWmsadgfjhiBJp30p’ # optional, overrides global Slack webhook 

Type: Budget

Define spend budgets and alert on budget overruns.

Required parameters:

  • type: budget
  • threshold: <amount> -- cost threshold in configured currency units
  • aggregation: <agg-parameter> -- configurable, accepts all aggregations supported by the aggregated cost model API
  • filter: <value> -- configurable, accepts a single filter value (comma-separated values unsupported)
  • window: <N>d or <M>h -- configurable, (1 ≤ N ≤ 7, 1 ≤ M ≤ 24)

Example Helm values.yaml:

			# Daily namespace budget alert on namespace `kubecost`
			- type: budget
				threshold: 50
				window: daily # or 1d
				aggregation: namespace
				filter: kubecost
			# 3d cluster budget alert on cluster `cluster-one`
			- type: budget
				threshold: 600
				window: 3d
				aggregation: cluster
				filter: cluster-one

Type: Spend Change

Detect unexpected spend increases/decreases relative to historical moving averages.

Required parameters:

  • type: spendChange
  • relativeThreshold: <N> -- configurable, N ≥ -1
  • aggregation: <agg-value> -- configurable, accepts all aggregations supported by the aggregated cost model API
  • window: <N>d or <M>h -- configurable, (1 ≤ N ≤ 7, 1 ≤ M ≤ 24)
  • baselineWindow: <N>d -- configurable, N ≥ 1

Optional parameters:

  • filter: <value> -- limit the aggregations that this alert will cover, accepts comma-separated values

Example Helm values.yaml:

			# Daily spend change alert on the 
			- type: spendChange
			  relativeThreshold: 0.20  # change relative to baseline average cost. Must be greater than -1 (can be negative).
			  window: 1d                # accepts ‘d’, ‘h’
			  baselineWindow: 30d       # previous window, offset by window
			  aggregation: namespace
			  filter: kubecost, default # accepts csv

Type: Kubecost Health Diagnostic

Enabling diagnostic alerts in Kubecost occur when an event impacts product uptime. This feature can be enabled in seconds from a values file. The following health events are detected:

  • Prometheus is unreachable
  • Kubecost metrics missing over last 5 minutes
  • More coming soon.

This alert only uses Slack (email coming soon), so it requires the globalSlackWebhookUrl field.

Example Helm values.yaml:

        # Kubecost Health Diagnostic
        - type: diagnostic
          window: 10m               

Versions Earlier than 1.79.0

This alert used to be configured via the notifications.alertConfigs.kubecostHealth flag seen here. If upgrading to version 1.79.0 or newer, remove the kubecostHealth flag, and append the alert definition shown above.

Type: Cluster Health

Cluster health alerts occur when the cluster health score changes by a specific threshold. The health score is calculated based on the following criteria:

  • Low Cluster Memory
  • Low Cluster CPU
  • Too Many Pods
  • Crash Looping Pods
  • Out of Memory Pods
  • Failed Jobs

This alert only uses Slack (email coming soon), so it requires the globalSlackWebhookUrl field.

Example Helm values.yaml:

        # Health Score Alert 
        - type: health              # Alerts when health score changes by a threshold
          window: 10m
          threshold: 5              # Send Alert if health scores changes by 5 or more

Alerts Scheduler

All times in UTC.

The back end scheduler runs if alertConfigs.enabled is set to true, and alerts at regular times of the day. Alert send times are determined by parsing the supplied window parameter. If alertConfigs.enabled is false, alerts are still configurable via the Kubecost UI Notifications tab, and will send alerts through the front end cron job scheduler.

Alert diagnostics with next and last scheduled run times are only available for alerts configured via Helm, and are viewed via <your-kubecost-url>/model/getCustomAlertDiagnostics.

Supported: weekly and daily special cases, <N>d, <M>h (1 ≤ N ≤ 7, 1 ≤ M ≤ 24) Currently Unsupported: time zone adjustments, windows greater than 7d, windows less than 1h

Scheduler Behavior

An <N>d alert sends at 00:00 UTC N day(s) from now, i.e., N days from now rounded down to midnight.

For example, a 5d alert scheduled on Monday will send on Saturday at 00:00, and subsequently the next Thursday at 00:00

An <N>h alert sends at the earliest time of day after now that is a multiple of N.

For example, a 6h alert scheduled at any time between 12 pm and 6 pm will send next at 6 pm and subsequently at 12 am the next day.

If 24 is not divisible by the hourly window, schedule at next multiple of <N>h after now, starting from the current day at 00:00.

For example, a 7h alert scheduled at 22:00 checks 00:00, 7:00, 14:00, and 21:00, before arriving at the next send time of 4:00 tomorrow.

Troubleshooting

Review these steps to verify alerts are being passed to the Kubecost application correctly.

First, ensure that the Helm values are successfully read into the configmap:

  • Confirm that the global.notifications.alertConfigs.enabled field is set to true
  • Run kubectl get configmap alert-configs -n kubecost -o json to view alerts configmap.
  • Ensure that the Helm values are successfully read into the configmap under alerts.json under the data field.
  • Example:
{
    "apiVersion": "v1",
    "data": {
        "alerts.json": "{\"alerts\":[{\"aggregation\":\"namespace\",\"efficiencyThreshold\":0.4,\"spendThreshold\":1,\"type\":\"efficiency\",\"window\":\"1d\"}],\"enabled\":true,\"frontendUrl\":\"http://localhost:9090\",\"globalAlertEmails\":[\"[email protected]\"],\"globalSlackWebhookUrl\":\"https://hooks.slack.com/services/TE6RTBNET/BFFK0P848/jFWms48dnxlk4BBPiBJp30p\",\"kubecostHealth\":true}"
    },
    "kind": "ConfigMap",
    "metadata": {
        "annotations": {
            "meta.helm.sh/release-name": "kubecost",
            "meta.helm.sh/release-namespace": "kubecost"
        },
        "creationTimestamp": "2021-01-13T04:14:52Z",
        "labels": {
            "app": "cost-analyzer",
            "app.kubernetes.io/instance": "kubecost-stage",
            "app.kubernetes.io/managed-by": "Helm",
	    ...
        },
        "name": "alert-configs",
        "namespace": "kubecost",
	...
    }
}

  • Ensure that the json string is successfully mapped to the appropriate configs

Next, confirm that Kubecost product has received configuration data:

  • Go to <your-kubecost-url>/notify.html in the Kubecost UI to view configured email and Slack settings, weekly updates, namespace updates, cluster budget, and namespace budget alerts.

Additionally, confirm that the alerts scheduler has properly parsed and scheduled a next run for each custom alert by visiting <your-kubecost-url>/model/getCustomAlertDiagnostics to view individual alert parameters as well as next and last scheduled run times for individual alerts.

  • Confirm that nextRun has been updated from "0001-01-01T00:00:00Z"
  • Settings that do not appear in the Notifications page (for example, spendChange alerts), but are visible from the /model/getCustomAlertDiagnostics endpoint are scheduled to send.

If nextRun fails to update, or alerts are not sending at the nextRun time, check pod logs by running kubectl logs $(kubectl get pods -n kubecost | awk '{print $1}' | grep "^kubecost-cost-analyzer.\{16\}") -n kubecost -c cost-model > kubecost-logs.txt

  • Common causes of misconfiguration include the following:
    • unsupported csv filters -- spendChange alerts accept filter as comma-separated values; other alert types do not.
    • unsupported alert type -- all alert type names are in camelCase -- check spelling and capitalization for all alert parameters
    • unsupported aggregation parameters -- see the aggregated cost model API for details

Have questions? Join our Slack community or contact us via email at [email protected]!

Edit this doc on Github