The helm installation makes it particularly simple to get going: https://github.com/kubernetes/charts/tree/master/stable/prometheus.
Now you just have to worry about
- Creating the alert
- Creating the alert rules
What kind of alert do you want and where do you want it to go? This is defined in the configMap for alertManager, which can be found in the alertmanagerFiles.alertmanager.yml portion of the values.yaml file of the helm chart.
alertmanagerFiles:
  alertmanager.yml:
    global: {}
      # slack_api_url: ''
    receivers:
      - name: default-receiver
        # slack_configs:
        #  - channel: '@you'
        #    send_resolved: true
    route:
      group_wait: 10s
      group_interval: 5m
      receiver: default-receiver
      repeat_interval: 3hAn example for creating an email alert may look something like:
global:
      smtp_smarthost: http://example-mail-service:25
      smtp_from: '[email protected]'
    receivers:
      - name: default-receiver
        email_configs:
        - to: [email protected]
    route:
      group_wait: 10s
      group_interval: 5m
      receiver: default-receiver
      repeat_interval: 3hWhen should alerts be fired? This is defined in the configMap for the server, which can be found in the serverFiles.alerts portion of the values.yaml file
serverFiles:
  alerts: {}An example for alerting when a HPA has scaled out to its max replicas:
serverFiles:
  alerts:
    groups:
    - name: group-name
      rules:
      - alert: MaxReplicaReached
        expr: kube_hpa_status_current_replicas{hpa="name-of-hpa"} == kube_hpa_spec_max_replicas{hpa="name-of-hpa"}
        for: 3m
        labels:
          severity: page
        annotations:
          summary: Trigger a alerts if max replicas is reached for more than 3 minutes