// Copyright 2013 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package rules import ( "context" "fmt" "math" "reflect" "strings" "testing" "time" "github.com/prometheus/common/model" "github.com/prometheus/prometheus/pkg/labels" "github.com/prometheus/prometheus/pkg/timestamp" "github.com/prometheus/prometheus/pkg/value" "github.com/prometheus/prometheus/promql" "github.com/prometheus/prometheus/storage" "github.com/prometheus/prometheus/util/testutil" ) func TestAlertingRule(t *testing.T) { suite, err := promql.NewTest(t, ` load 5m http_requests{job="app-server", instance="0", group="canary", severity="overwrite-me"} 75 85 95 105 105 95 85 http_requests{job="app-server", instance="1", group="canary", severity="overwrite-me"} 80 90 100 110 120 130 140 `) if err != nil { t.Fatal(err) } defer suite.Close() if err := suite.Run(); err != nil { t.Fatal(err) } expr, err := promql.ParseExpr(`http_requests{group="canary", job="app-server"} < 100`) if err != nil { t.Fatalf("Unable to parse alert expression: %s", err) } rule := NewAlertingRule( "HTTPRequestRateLow", expr, time.Minute, labels.FromStrings("severity", "{{\"c\"}}ritical"), nil, ) baseTime := time.Unix(0, 0) var tests = []struct { time time.Duration result []string }{ { time: 0, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="1", job="app-server", severity="critical"} => 1 @[%v]`, }, }, { time: 5 * time.Minute, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 0 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="1", job="app-server", severity="critical"} => 0 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="1", job="app-server", severity="critical"} => 1 @[%v]`, }, }, { time: 10 * time.Minute, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="1", job="app-server", severity="critical"} => 0 @[%v]`, }, }, { time: 15 * time.Minute, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 0 @[%v]`, }, }, { time: 20 * time.Minute, result: []string{}, }, { time: 25 * time.Minute, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`, }, }, { time: 30 * time.Minute, result: []string{ `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="pending", group="canary", instance="0", job="app-server", severity="critical"} => 0 @[%v]`, `{__name__="ALERTS", alertname="HTTPRequestRateLow", alertstate="firing", group="canary", instance="0", job="app-server", severity="critical"} => 1 @[%v]`, }, }, } for i, test := range tests { evalTime := baseTime.Add(test.time) res, err := rule.Eval(suite.Context(), evalTime, suite.QueryEngine(), nil) if err != nil { t.Fatalf("Error during alerting rule evaluation: %s", err) } actual := strings.Split(res.String(), "\n") expected := annotateWithTime(test.result, evalTime) if actual[0] == "" { actual = []string{} } if len(actual) != len(expected) { t.Errorf("%d. Number of samples in expected and actual output don't match (%d vs. %d)", i, len(expected), len(actual)) } for j, expectedSample := range expected { found := false for _, actualSample := range actual { if actualSample == expectedSample { found = true } } if !found { t.Errorf("%d.%d. Couldn't find expected sample in output: '%v'", i, j, expectedSample) } } if t.Failed() { t.Errorf("%d. Expected and actual outputs don't match:", i) t.Fatalf("Expected:\n%v\n----\nActual:\n%v", strings.Join(expected, "\n"), strings.Join(actual, "\n")) } for _, aa := range rule.ActiveAlerts() { if v := aa.Labels.Get(model.MetricNameLabel); v != "" { t.Fatalf("%s label set on active alert: %s", model.MetricNameLabel, aa.Labels) } } } } func annotateWithTime(lines []string, ts time.Time) []string { annotatedLines := []string{} for _, line := range lines { annotatedLines = append(annotatedLines, fmt.Sprintf(line, timestamp.FromTime(ts))) } return annotatedLines } func TestStaleness(t *testing.T) { storage := testutil.NewStorage(t) defer storage.Close() engine := promql.NewEngine(storage, nil) opts := &ManagerOptions{ QueryEngine: engine, Appendable: storage, Context: context.Background(), } expr, err := promql.ParseExpr("a + 1") if err != nil { t.Fatal(err) } rule := NewRecordingRule("a_plus_one", expr, labels.Labels{}) group := NewGroup("default", time.Second, []Rule{rule}, opts) // A time series that has two samples and then goes stale. app, _ := storage.Appender() app.Add(labels.FromStrings(model.MetricNameLabel, "a"), 0, 1) app.Add(labels.FromStrings(model.MetricNameLabel, "a"), 1000, 2) app.Add(labels.FromStrings(model.MetricNameLabel, "a"), 2000, math.Float64frombits(value.StaleNaN)) if err = app.Commit(); err != nil { t.Fatal(err) } // Execute 3 times, 1 second apart. group.Eval(time.Unix(0, 0)) group.Eval(time.Unix(1, 0)) group.Eval(time.Unix(2, 0)) querier, err := storage.Querier(0, 2000) defer querier.Close() if err != nil { t.Fatal(err) } matcher, _ := labels.NewMatcher(labels.MatchEqual, model.MetricNameLabel, "a_plus_one") samples, err := readSeriesSet(querier.Select(matcher)) if err != nil { t.Fatal(err) } metric := labels.FromStrings(model.MetricNameLabel, "a_plus_one").String() metricSample, ok := samples[metric] if !ok { t.Fatalf("Series %s not returned.", metric) } if !value.IsStaleNaN(metricSample[2].V) { t.Fatalf("Appended second sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(metricSample[2].V)) } metricSample[2].V = 42 // reflect.DeepEqual cannot handle NaN. want := map[string][]promql.Point{ metric: []promql.Point{{0, 2}, {1000, 3}, {2000, 42}}, } if !reflect.DeepEqual(want, samples) { t.Fatalf("Returned samples not as expected. Wanted: %+v Got: %+v", want, samples) } } // Convert a SeriesSet into a form useable with reflect.DeepEqual. func readSeriesSet(ss storage.SeriesSet) (map[string][]promql.Point, error) { result := map[string][]promql.Point{} for ss.Next() { series := ss.At() points := []promql.Point{} it := series.Iterator() for it.Next() { t, v := it.At() points = append(points, promql.Point{T: t, V: v}) } name := series.Labels().String() result[name] = points } return result, ss.Err() } func TestCopyState(t *testing.T) { oldGroup := &Group{ rules: []Rule{ NewAlertingRule("alert", nil, 0, nil, nil), NewRecordingRule("rule1", nil, nil), NewRecordingRule("rule2", nil, nil), NewRecordingRule("rule3", nil, nil), NewRecordingRule("rule3", nil, nil), }, seriesInPreviousEval: []map[string]labels.Labels{ map[string]labels.Labels{"a": nil}, map[string]labels.Labels{"r1": nil}, map[string]labels.Labels{"r2": nil}, map[string]labels.Labels{"r3a": nil}, map[string]labels.Labels{"r3b": nil}, }, } oldGroup.rules[0].(*AlertingRule).active[42] = nil newGroup := &Group{ rules: []Rule{ NewRecordingRule("rule3", nil, nil), NewRecordingRule("rule3", nil, nil), NewRecordingRule("rule3", nil, nil), NewAlertingRule("alert", nil, 0, nil, nil), NewRecordingRule("rule1", nil, nil), NewRecordingRule("rule4", nil, nil), }, seriesInPreviousEval: make([]map[string]labels.Labels, 6), } newGroup.copyState(oldGroup) want := []map[string]labels.Labels{ map[string]labels.Labels{"r3a": nil}, map[string]labels.Labels{"r3b": nil}, nil, map[string]labels.Labels{"a": nil}, map[string]labels.Labels{"r1": nil}, nil, } if !reflect.DeepEqual(want, newGroup.seriesInPreviousEval) { t.Fatalf("seriesInPreviousEval not as expected. Wanted: %+v Got: %+v", want, newGroup.seriesInPreviousEval) } if !reflect.DeepEqual(oldGroup.rules[0], newGroup.rules[3]) { t.Fatalf("Active alerts not as expected. Wanted: %+v Got: %+v", oldGroup.rules[0], oldGroup.rules[3]) } }