708 lines
22 KiB
Go
708 lines
22 KiB
Go
// Copyright 2018 Prometheus Team
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package v2
|
|
|
|
import (
|
|
"fmt"
|
|
"net/http"
|
|
"regexp"
|
|
"sort"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/go-kit/log"
|
|
"github.com/go-kit/log/level"
|
|
"github.com/go-openapi/analysis"
|
|
"github.com/go-openapi/loads"
|
|
"github.com/go-openapi/runtime/middleware"
|
|
"github.com/go-openapi/strfmt"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
prometheus_model "github.com/prometheus/common/model"
|
|
"github.com/prometheus/common/version"
|
|
"github.com/rs/cors"
|
|
|
|
"github.com/prometheus/alertmanager/api/metrics"
|
|
open_api_models "github.com/prometheus/alertmanager/api/v2/models"
|
|
"github.com/prometheus/alertmanager/api/v2/restapi"
|
|
"github.com/prometheus/alertmanager/api/v2/restapi/operations"
|
|
alert_ops "github.com/prometheus/alertmanager/api/v2/restapi/operations/alert"
|
|
alertgroup_ops "github.com/prometheus/alertmanager/api/v2/restapi/operations/alertgroup"
|
|
general_ops "github.com/prometheus/alertmanager/api/v2/restapi/operations/general"
|
|
receiver_ops "github.com/prometheus/alertmanager/api/v2/restapi/operations/receiver"
|
|
silence_ops "github.com/prometheus/alertmanager/api/v2/restapi/operations/silence"
|
|
"github.com/prometheus/alertmanager/cluster"
|
|
"github.com/prometheus/alertmanager/config"
|
|
"github.com/prometheus/alertmanager/dispatch"
|
|
"github.com/prometheus/alertmanager/pkg/labels"
|
|
"github.com/prometheus/alertmanager/provider"
|
|
"github.com/prometheus/alertmanager/silence"
|
|
"github.com/prometheus/alertmanager/silence/silencepb"
|
|
"github.com/prometheus/alertmanager/types"
|
|
)
|
|
|
|
// API represents an Alertmanager API v2
|
|
type API struct {
|
|
peer cluster.ClusterPeer
|
|
silences *silence.Silences
|
|
alerts provider.Alerts
|
|
alertGroups groupsFn
|
|
getAlertStatus getAlertStatusFn
|
|
uptime time.Time
|
|
|
|
// mtx protects alertmanagerConfig, setAlertStatus and route.
|
|
mtx sync.RWMutex
|
|
// resolveTimeout represents the default resolve timeout that an alert is
|
|
// assigned if no end time is specified.
|
|
alertmanagerConfig *config.Config
|
|
route *dispatch.Route
|
|
setAlertStatus setAlertStatusFn
|
|
|
|
logger log.Logger
|
|
m *metrics.Alerts
|
|
|
|
Handler http.Handler
|
|
}
|
|
|
|
type (
|
|
groupsFn func(func(*dispatch.Route) bool, func(*types.Alert, time.Time) bool) (dispatch.AlertGroups, map[prometheus_model.Fingerprint][]string)
|
|
getAlertStatusFn func(prometheus_model.Fingerprint) types.AlertStatus
|
|
setAlertStatusFn func(prometheus_model.LabelSet)
|
|
)
|
|
|
|
// NewAPI returns a new Alertmanager API v2
|
|
func NewAPI(
|
|
alerts provider.Alerts,
|
|
gf groupsFn,
|
|
sf getAlertStatusFn,
|
|
silences *silence.Silences,
|
|
peer cluster.ClusterPeer,
|
|
l log.Logger,
|
|
r prometheus.Registerer,
|
|
) (*API, error) {
|
|
api := API{
|
|
alerts: alerts,
|
|
getAlertStatus: sf,
|
|
alertGroups: gf,
|
|
peer: peer,
|
|
silences: silences,
|
|
logger: l,
|
|
m: metrics.NewAlerts("v2", r),
|
|
uptime: time.Now(),
|
|
}
|
|
|
|
// Load embedded swagger file.
|
|
swaggerSpec, swaggerSpecAnalysis, err := getSwaggerSpec()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Create new service API.
|
|
openAPI := operations.NewAlertmanagerAPI(swaggerSpec)
|
|
|
|
// Skip the redoc middleware, only serving the OpenAPI specification and
|
|
// the API itself via RoutesHandler. See:
|
|
// https://github.com/go-swagger/go-swagger/issues/1779
|
|
openAPI.Middleware = func(b middleware.Builder) http.Handler {
|
|
// Manually create the context so that we can use the singleton swaggerSpecAnalysis.
|
|
swaggerContext := middleware.NewRoutableContextWithAnalyzedSpec(swaggerSpec, swaggerSpecAnalysis, openAPI, nil)
|
|
return middleware.Spec("", swaggerSpec.Raw(), swaggerContext.RoutesHandler(b))
|
|
}
|
|
|
|
openAPI.AlertGetAlertsHandler = alert_ops.GetAlertsHandlerFunc(api.getAlertsHandler)
|
|
openAPI.AlertPostAlertsHandler = alert_ops.PostAlertsHandlerFunc(api.postAlertsHandler)
|
|
openAPI.AlertgroupGetAlertGroupsHandler = alertgroup_ops.GetAlertGroupsHandlerFunc(api.getAlertGroupsHandler)
|
|
openAPI.GeneralGetStatusHandler = general_ops.GetStatusHandlerFunc(api.getStatusHandler)
|
|
openAPI.ReceiverGetReceiversHandler = receiver_ops.GetReceiversHandlerFunc(api.getReceiversHandler)
|
|
openAPI.SilenceDeleteSilenceHandler = silence_ops.DeleteSilenceHandlerFunc(api.deleteSilenceHandler)
|
|
openAPI.SilenceGetSilenceHandler = silence_ops.GetSilenceHandlerFunc(api.getSilenceHandler)
|
|
openAPI.SilenceGetSilencesHandler = silence_ops.GetSilencesHandlerFunc(api.getSilencesHandler)
|
|
openAPI.SilencePostSilencesHandler = silence_ops.PostSilencesHandlerFunc(api.postSilencesHandler)
|
|
|
|
handleCORS := cors.Default().Handler
|
|
api.Handler = handleCORS(openAPI.Serve(nil))
|
|
|
|
return &api, nil
|
|
}
|
|
|
|
func (api *API) requestLogger(req *http.Request) log.Logger {
|
|
return log.With(api.logger, "path", req.URL.Path, "method", req.Method)
|
|
}
|
|
|
|
// Update sets the API struct members that may change between reloads of alertmanager.
|
|
func (api *API) Update(cfg *config.Config, setAlertStatus setAlertStatusFn) {
|
|
api.mtx.Lock()
|
|
defer api.mtx.Unlock()
|
|
|
|
api.alertmanagerConfig = cfg
|
|
api.route = dispatch.NewRoute(cfg.Route, nil)
|
|
api.setAlertStatus = setAlertStatus
|
|
}
|
|
|
|
func (api *API) getStatusHandler(params general_ops.GetStatusParams) middleware.Responder {
|
|
api.mtx.RLock()
|
|
defer api.mtx.RUnlock()
|
|
|
|
original := api.alertmanagerConfig.String()
|
|
uptime := strfmt.DateTime(api.uptime)
|
|
|
|
status := open_api_models.ClusterStatusStatusDisabled
|
|
|
|
resp := open_api_models.AlertmanagerStatus{
|
|
Uptime: &uptime,
|
|
VersionInfo: &open_api_models.VersionInfo{
|
|
Version: &version.Version,
|
|
Revision: &version.Revision,
|
|
Branch: &version.Branch,
|
|
BuildUser: &version.BuildUser,
|
|
BuildDate: &version.BuildDate,
|
|
GoVersion: &version.GoVersion,
|
|
},
|
|
Config: &open_api_models.AlertmanagerConfig{
|
|
Original: &original,
|
|
},
|
|
Cluster: &open_api_models.ClusterStatus{
|
|
Status: &status,
|
|
Peers: []*open_api_models.PeerStatus{},
|
|
},
|
|
}
|
|
|
|
// If alertmanager cluster feature is disabled, then api.peers == nil.
|
|
if api.peer != nil {
|
|
status := api.peer.Status()
|
|
|
|
peers := []*open_api_models.PeerStatus{}
|
|
for _, n := range api.peer.Peers() {
|
|
address := n.Address()
|
|
name := n.Name()
|
|
peers = append(peers, &open_api_models.PeerStatus{
|
|
Name: &name,
|
|
Address: &address,
|
|
})
|
|
}
|
|
|
|
sort.Slice(peers, func(i, j int) bool {
|
|
return *peers[i].Name < *peers[j].Name
|
|
})
|
|
|
|
resp.Cluster = &open_api_models.ClusterStatus{
|
|
Name: api.peer.Name(),
|
|
Status: &status,
|
|
Peers: peers,
|
|
}
|
|
}
|
|
|
|
return general_ops.NewGetStatusOK().WithPayload(&resp)
|
|
}
|
|
|
|
func (api *API) getReceiversHandler(params receiver_ops.GetReceiversParams) middleware.Responder {
|
|
api.mtx.RLock()
|
|
defer api.mtx.RUnlock()
|
|
|
|
receivers := make([]*open_api_models.Receiver, 0, len(api.alertmanagerConfig.Receivers))
|
|
for _, r := range api.alertmanagerConfig.Receivers {
|
|
receivers = append(receivers, &open_api_models.Receiver{Name: &r.Name})
|
|
}
|
|
|
|
return receiver_ops.NewGetReceiversOK().WithPayload(receivers)
|
|
}
|
|
|
|
func (api *API) getAlertsHandler(params alert_ops.GetAlertsParams) middleware.Responder {
|
|
var (
|
|
receiverFilter *regexp.Regexp
|
|
// Initialize result slice to prevent api returning `null` when there
|
|
// are no alerts present
|
|
res = open_api_models.GettableAlerts{}
|
|
ctx = params.HTTPRequest.Context()
|
|
|
|
logger = api.requestLogger(params.HTTPRequest)
|
|
)
|
|
|
|
matchers, err := parseFilter(params.Filter)
|
|
if err != nil {
|
|
level.Debug(logger).Log("msg", "Failed to parse matchers", "err", err)
|
|
return alertgroup_ops.NewGetAlertGroupsBadRequest().WithPayload(err.Error())
|
|
}
|
|
|
|
if params.Receiver != nil {
|
|
receiverFilter, err = regexp.Compile("^(?:" + *params.Receiver + ")$")
|
|
if err != nil {
|
|
level.Debug(logger).Log("msg", "Failed to compile receiver regex", "err", err)
|
|
return alert_ops.
|
|
NewGetAlertsBadRequest().
|
|
WithPayload(
|
|
fmt.Sprintf("failed to parse receiver param: %v", err.Error()),
|
|
)
|
|
}
|
|
}
|
|
|
|
alerts := api.alerts.GetPending()
|
|
defer alerts.Close()
|
|
|
|
alertFilter := api.alertFilter(matchers, *params.Silenced, *params.Inhibited, *params.Active)
|
|
now := time.Now()
|
|
|
|
api.mtx.RLock()
|
|
for a := range alerts.Next() {
|
|
if err = alerts.Err(); err != nil {
|
|
break
|
|
}
|
|
if err = ctx.Err(); err != nil {
|
|
break
|
|
}
|
|
|
|
routes := api.route.Match(a.Labels)
|
|
receivers := make([]string, 0, len(routes))
|
|
for _, r := range routes {
|
|
receivers = append(receivers, r.RouteOpts.Receiver)
|
|
}
|
|
|
|
if receiverFilter != nil && !receiversMatchFilter(receivers, receiverFilter) {
|
|
continue
|
|
}
|
|
|
|
if !alertFilter(a, now) {
|
|
continue
|
|
}
|
|
|
|
alert := AlertToOpenAPIAlert(a, api.getAlertStatus(a.Fingerprint()), receivers)
|
|
|
|
res = append(res, alert)
|
|
}
|
|
api.mtx.RUnlock()
|
|
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to get alerts", "err", err)
|
|
return alert_ops.NewGetAlertsInternalServerError().WithPayload(err.Error())
|
|
}
|
|
sort.Slice(res, func(i, j int) bool {
|
|
return *res[i].Fingerprint < *res[j].Fingerprint
|
|
})
|
|
|
|
return alert_ops.NewGetAlertsOK().WithPayload(res)
|
|
}
|
|
|
|
func (api *API) postAlertsHandler(params alert_ops.PostAlertsParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
alerts := OpenAPIAlertsToAlerts(params.Alerts)
|
|
now := time.Now()
|
|
|
|
api.mtx.RLock()
|
|
resolveTimeout := time.Duration(api.alertmanagerConfig.Global.ResolveTimeout)
|
|
api.mtx.RUnlock()
|
|
|
|
for _, alert := range alerts {
|
|
alert.UpdatedAt = now
|
|
|
|
// Ensure StartsAt is set.
|
|
if alert.StartsAt.IsZero() {
|
|
if alert.EndsAt.IsZero() {
|
|
alert.StartsAt = now
|
|
} else {
|
|
alert.StartsAt = alert.EndsAt
|
|
}
|
|
}
|
|
// If no end time is defined, set a timeout after which an alert
|
|
// is marked resolved if it is not updated.
|
|
if alert.EndsAt.IsZero() {
|
|
alert.Timeout = true
|
|
alert.EndsAt = now.Add(resolveTimeout)
|
|
}
|
|
if alert.EndsAt.After(time.Now()) {
|
|
api.m.Firing().Inc()
|
|
} else {
|
|
api.m.Resolved().Inc()
|
|
}
|
|
}
|
|
|
|
// Make a best effort to insert all alerts that are valid.
|
|
var (
|
|
validAlerts = make([]*types.Alert, 0, len(alerts))
|
|
validationErrs = &types.MultiError{}
|
|
)
|
|
for _, a := range alerts {
|
|
removeEmptyLabels(a.Labels)
|
|
|
|
if err := a.Validate(); err != nil {
|
|
validationErrs.Add(err)
|
|
api.m.Invalid().Inc()
|
|
continue
|
|
}
|
|
validAlerts = append(validAlerts, a)
|
|
}
|
|
if err := api.alerts.Put(validAlerts...); err != nil {
|
|
level.Error(logger).Log("msg", "Failed to create alerts", "err", err)
|
|
return alert_ops.NewPostAlertsInternalServerError().WithPayload(err.Error())
|
|
}
|
|
|
|
if validationErrs.Len() > 0 {
|
|
level.Error(logger).Log("msg", "Failed to validate alerts", "err", validationErrs.Error())
|
|
return alert_ops.NewPostAlertsBadRequest().WithPayload(validationErrs.Error())
|
|
}
|
|
|
|
return alert_ops.NewPostAlertsOK()
|
|
}
|
|
|
|
func (api *API) getAlertGroupsHandler(params alertgroup_ops.GetAlertGroupsParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
matchers, err := parseFilter(params.Filter)
|
|
if err != nil {
|
|
level.Debug(logger).Log("msg", "Failed to parse matchers", "err", err)
|
|
return alertgroup_ops.NewGetAlertGroupsBadRequest().WithPayload(err.Error())
|
|
}
|
|
|
|
var receiverFilter *regexp.Regexp
|
|
if params.Receiver != nil {
|
|
receiverFilter, err = regexp.Compile("^(?:" + *params.Receiver + ")$")
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to compile receiver regex", "err", err)
|
|
return alertgroup_ops.
|
|
NewGetAlertGroupsBadRequest().
|
|
WithPayload(
|
|
fmt.Sprintf("failed to parse receiver param: %v", err.Error()),
|
|
)
|
|
}
|
|
}
|
|
|
|
rf := func(receiverFilter *regexp.Regexp) func(r *dispatch.Route) bool {
|
|
return func(r *dispatch.Route) bool {
|
|
receiver := r.RouteOpts.Receiver
|
|
if receiverFilter != nil && !receiverFilter.MatchString(receiver) {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
}(receiverFilter)
|
|
|
|
af := api.alertFilter(matchers, *params.Silenced, *params.Inhibited, *params.Active)
|
|
alertGroups, allReceivers := api.alertGroups(rf, af)
|
|
|
|
res := make(open_api_models.AlertGroups, 0, len(alertGroups))
|
|
|
|
for _, alertGroup := range alertGroups {
|
|
ag := &open_api_models.AlertGroup{
|
|
Receiver: &open_api_models.Receiver{Name: &alertGroup.Receiver},
|
|
Labels: ModelLabelSetToAPILabelSet(alertGroup.Labels),
|
|
Alerts: make([]*open_api_models.GettableAlert, 0, len(alertGroup.Alerts)),
|
|
}
|
|
|
|
for _, alert := range alertGroup.Alerts {
|
|
fp := alert.Fingerprint()
|
|
receivers := allReceivers[fp]
|
|
status := api.getAlertStatus(fp)
|
|
apiAlert := AlertToOpenAPIAlert(alert, status, receivers)
|
|
ag.Alerts = append(ag.Alerts, apiAlert)
|
|
}
|
|
res = append(res, ag)
|
|
}
|
|
|
|
return alertgroup_ops.NewGetAlertGroupsOK().WithPayload(res)
|
|
}
|
|
|
|
func (api *API) alertFilter(matchers []*labels.Matcher, silenced, inhibited, active bool) func(a *types.Alert, now time.Time) bool {
|
|
return func(a *types.Alert, now time.Time) bool {
|
|
if !a.EndsAt.IsZero() && a.EndsAt.Before(now) {
|
|
return false
|
|
}
|
|
|
|
// Set alert's current status based on its label set.
|
|
api.setAlertStatus(a.Labels)
|
|
|
|
// Get alert's current status after seeing if it is suppressed.
|
|
status := api.getAlertStatus(a.Fingerprint())
|
|
|
|
if !active && status.State == types.AlertStateActive {
|
|
return false
|
|
}
|
|
|
|
if !silenced && len(status.SilencedBy) != 0 {
|
|
return false
|
|
}
|
|
|
|
if !inhibited && len(status.InhibitedBy) != 0 {
|
|
return false
|
|
}
|
|
|
|
return alertMatchesFilterLabels(&a.Alert, matchers)
|
|
}
|
|
}
|
|
|
|
func removeEmptyLabels(ls prometheus_model.LabelSet) {
|
|
for k, v := range ls {
|
|
if string(v) == "" {
|
|
delete(ls, k)
|
|
}
|
|
}
|
|
}
|
|
|
|
func receiversMatchFilter(receivers []string, filter *regexp.Regexp) bool {
|
|
for _, r := range receivers {
|
|
if filter.MatchString(r) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
func alertMatchesFilterLabels(a *prometheus_model.Alert, matchers []*labels.Matcher) bool {
|
|
sms := make(map[string]string)
|
|
for name, value := range a.Labels {
|
|
sms[string(name)] = string(value)
|
|
}
|
|
return matchFilterLabels(matchers, sms)
|
|
}
|
|
|
|
func matchFilterLabels(matchers []*labels.Matcher, sms map[string]string) bool {
|
|
for _, m := range matchers {
|
|
v, prs := sms[m.Name]
|
|
switch m.Type {
|
|
case labels.MatchNotRegexp, labels.MatchNotEqual:
|
|
if m.Value == "" && prs {
|
|
continue
|
|
}
|
|
if !m.Matches(v) {
|
|
return false
|
|
}
|
|
default:
|
|
if m.Value == "" && !prs {
|
|
continue
|
|
}
|
|
if !m.Matches(v) {
|
|
return false
|
|
}
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
func (api *API) getSilencesHandler(params silence_ops.GetSilencesParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
matchers := []*labels.Matcher{}
|
|
if params.Filter != nil {
|
|
for _, matcherString := range params.Filter {
|
|
matcher, err := labels.ParseMatcher(matcherString)
|
|
if err != nil {
|
|
level.Debug(logger).Log("msg", "Failed to parse matchers", "err", err)
|
|
return alert_ops.NewGetAlertsBadRequest().WithPayload(err.Error())
|
|
}
|
|
|
|
matchers = append(matchers, matcher)
|
|
}
|
|
}
|
|
|
|
psils, _, err := api.silences.Query()
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to get silences", "err", err)
|
|
return silence_ops.NewGetSilencesInternalServerError().WithPayload(err.Error())
|
|
}
|
|
|
|
sils := open_api_models.GettableSilences{}
|
|
for _, ps := range psils {
|
|
if !CheckSilenceMatchesFilterLabels(ps, matchers) {
|
|
continue
|
|
}
|
|
silence, err := GettableSilenceFromProto(ps)
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to unmarshal silence from proto", "err", err)
|
|
return silence_ops.NewGetSilencesInternalServerError().WithPayload(err.Error())
|
|
}
|
|
sils = append(sils, &silence)
|
|
}
|
|
|
|
SortSilences(sils)
|
|
|
|
return silence_ops.NewGetSilencesOK().WithPayload(sils)
|
|
}
|
|
|
|
var silenceStateOrder = map[types.SilenceState]int{
|
|
types.SilenceStateActive: 1,
|
|
types.SilenceStatePending: 2,
|
|
types.SilenceStateExpired: 3,
|
|
}
|
|
|
|
// SortSilences sorts first according to the state "active, pending, expired"
|
|
// then by end time or start time depending on the state.
|
|
// active silences should show the next to expire first
|
|
// pending silences are ordered based on which one starts next
|
|
// expired are ordered based on which one expired most recently
|
|
func SortSilences(sils open_api_models.GettableSilences) {
|
|
sort.Slice(sils, func(i, j int) bool {
|
|
state1 := types.SilenceState(*sils[i].Status.State)
|
|
state2 := types.SilenceState(*sils[j].Status.State)
|
|
if state1 != state2 {
|
|
return silenceStateOrder[state1] < silenceStateOrder[state2]
|
|
}
|
|
switch state1 {
|
|
case types.SilenceStateActive:
|
|
endsAt1 := time.Time(*sils[i].Silence.EndsAt)
|
|
endsAt2 := time.Time(*sils[j].Silence.EndsAt)
|
|
return endsAt1.Before(endsAt2)
|
|
case types.SilenceStatePending:
|
|
startsAt1 := time.Time(*sils[i].Silence.StartsAt)
|
|
startsAt2 := time.Time(*sils[j].Silence.StartsAt)
|
|
return startsAt1.Before(startsAt2)
|
|
case types.SilenceStateExpired:
|
|
endsAt1 := time.Time(*sils[i].Silence.EndsAt)
|
|
endsAt2 := time.Time(*sils[j].Silence.EndsAt)
|
|
return endsAt1.After(endsAt2)
|
|
}
|
|
return false
|
|
})
|
|
}
|
|
|
|
// CheckSilenceMatchesFilterLabels returns true if
|
|
// a given silence matches a list of matchers.
|
|
// A silence matches a filter (list of matchers) if
|
|
// for all matchers in the filter, there exists a matcher in the silence
|
|
// such that their names, types, and values are equivalent.
|
|
func CheckSilenceMatchesFilterLabels(s *silencepb.Silence, matchers []*labels.Matcher) bool {
|
|
for _, matcher := range matchers {
|
|
found := false
|
|
for _, m := range s.Matchers {
|
|
if matcher.Name == m.Name &&
|
|
(matcher.Type == labels.MatchEqual && m.Type == silencepb.Matcher_EQUAL ||
|
|
matcher.Type == labels.MatchRegexp && m.Type == silencepb.Matcher_REGEXP ||
|
|
matcher.Type == labels.MatchNotEqual && m.Type == silencepb.Matcher_NOT_EQUAL ||
|
|
matcher.Type == labels.MatchNotRegexp && m.Type == silencepb.Matcher_NOT_REGEXP) &&
|
|
matcher.Value == m.Pattern {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
if !found {
|
|
return false
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
func (api *API) getSilenceHandler(params silence_ops.GetSilenceParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
sils, _, err := api.silences.Query(silence.QIDs(params.SilenceID.String()))
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to get silence by id", "err", err, "id", params.SilenceID.String())
|
|
return silence_ops.NewGetSilenceInternalServerError().WithPayload(err.Error())
|
|
}
|
|
|
|
if len(sils) == 0 {
|
|
level.Error(logger).Log("msg", "Failed to find silence", "err", err, "id", params.SilenceID.String())
|
|
return silence_ops.NewGetSilenceNotFound()
|
|
}
|
|
|
|
sil, err := GettableSilenceFromProto(sils[0])
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to convert unmarshal from proto", "err", err)
|
|
return silence_ops.NewGetSilenceInternalServerError().WithPayload(err.Error())
|
|
}
|
|
|
|
return silence_ops.NewGetSilenceOK().WithPayload(&sil)
|
|
}
|
|
|
|
func (api *API) deleteSilenceHandler(params silence_ops.DeleteSilenceParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
sid := params.SilenceID.String()
|
|
if err := api.silences.Expire(sid); err != nil {
|
|
level.Error(logger).Log("msg", "Failed to expire silence", "err", err)
|
|
return silence_ops.NewDeleteSilenceInternalServerError().WithPayload(err.Error())
|
|
}
|
|
return silence_ops.NewDeleteSilenceOK()
|
|
}
|
|
|
|
func (api *API) postSilencesHandler(params silence_ops.PostSilencesParams) middleware.Responder {
|
|
logger := api.requestLogger(params.HTTPRequest)
|
|
|
|
sil, err := PostableSilenceToProto(params.Silence)
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to marshal silence to proto", "err", err)
|
|
return silence_ops.NewPostSilencesBadRequest().WithPayload(
|
|
fmt.Sprintf("failed to convert API silence to internal silence: %v", err.Error()),
|
|
)
|
|
}
|
|
|
|
if sil.StartsAt.After(sil.EndsAt) || sil.StartsAt.Equal(sil.EndsAt) {
|
|
msg := "Failed to create silence: start time must be before end time"
|
|
level.Error(logger).Log("msg", msg, "starts_at", sil.StartsAt, "ends_at", sil.EndsAt)
|
|
return silence_ops.NewPostSilencesBadRequest().WithPayload(msg)
|
|
}
|
|
|
|
if sil.EndsAt.Before(time.Now()) {
|
|
msg := "Failed to create silence: end time can't be in the past"
|
|
level.Error(logger).Log("msg", msg, "ends_at", sil.EndsAt)
|
|
return silence_ops.NewPostSilencesBadRequest().WithPayload(msg)
|
|
}
|
|
|
|
sid, err := api.silences.Set(sil)
|
|
if err != nil {
|
|
level.Error(logger).Log("msg", "Failed to create silence", "err", err)
|
|
if err == silence.ErrNotFound {
|
|
return silence_ops.NewPostSilencesNotFound().WithPayload(err.Error())
|
|
}
|
|
return silence_ops.NewPostSilencesBadRequest().WithPayload(err.Error())
|
|
}
|
|
|
|
return silence_ops.NewPostSilencesOK().WithPayload(&silence_ops.PostSilencesOKBody{
|
|
SilenceID: sid,
|
|
})
|
|
}
|
|
|
|
func parseFilter(filter []string) ([]*labels.Matcher, error) {
|
|
matchers := make([]*labels.Matcher, 0, len(filter))
|
|
for _, matcherString := range filter {
|
|
matcher, err := labels.ParseMatcher(matcherString)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
matchers = append(matchers, matcher)
|
|
}
|
|
return matchers, nil
|
|
}
|
|
|
|
var (
|
|
swaggerSpecCacheMx sync.Mutex
|
|
swaggerSpecCache *loads.Document
|
|
swaggerSpecAnalysisCache *analysis.Spec
|
|
)
|
|
|
|
// getSwaggerSpec loads and caches the swagger spec. If a cached version already exists,
|
|
// it returns the cached one. The reason why we cache it is because some downstream projects
|
|
// (e.g. Grafana Mimir) creates many Alertmanager instances in the same process, so they would
|
|
// incur in a significant memory penalty if we would reload the swagger spec each time.
|
|
func getSwaggerSpec() (*loads.Document, *analysis.Spec, error) {
|
|
swaggerSpecCacheMx.Lock()
|
|
defer swaggerSpecCacheMx.Unlock()
|
|
|
|
// Check if a cached version exists.
|
|
if swaggerSpecCache != nil {
|
|
return swaggerSpecCache, swaggerSpecAnalysisCache, nil
|
|
}
|
|
|
|
// Load embedded swagger file.
|
|
swaggerSpec, err := loads.Analyzed(restapi.SwaggerJSON, "")
|
|
if err != nil {
|
|
return nil, nil, fmt.Errorf("failed to load embedded swagger file: %w", err)
|
|
}
|
|
|
|
swaggerSpecCache = swaggerSpec
|
|
swaggerSpecAnalysisCache = analysis.New(swaggerSpec.Spec())
|
|
return swaggerSpec, swaggerSpecAnalysisCache, nil
|
|
}
|