mirror of
https://github.com/prometheus/alertmanager
synced 2024-12-27 08:32:15 +00:00
c39a913f8a
This change enables race detection when running the tests. It also fixes a couple of existing race conditions.
173 lines
3.7 KiB
Go
173 lines
3.7 KiB
Go
// Copyright 2015 Prometheus Team
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package test
|
|
|
|
import (
|
|
"fmt"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/prometheus/common/model"
|
|
)
|
|
|
|
// Collector gathers alerts received by a notification receiver
|
|
// and verifies whether all arrived and within the correct time boundaries.
|
|
type Collector struct {
|
|
t *testing.T
|
|
name string
|
|
opts *AcceptanceOpts
|
|
|
|
collected map[float64][]model.Alerts
|
|
expected map[Interval][]model.Alerts
|
|
|
|
mtx sync.RWMutex
|
|
}
|
|
|
|
func (c *Collector) String() string {
|
|
return c.name
|
|
}
|
|
|
|
func batchesEqual(as, bs model.Alerts, opts *AcceptanceOpts) bool {
|
|
if len(as) != len(bs) {
|
|
return false
|
|
}
|
|
|
|
for _, a := range as {
|
|
found := false
|
|
for _, b := range bs {
|
|
if equalAlerts(a, b, opts) {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
if !found {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
// latest returns the latest relative point in time where a notification is
|
|
// expected.
|
|
func (c *Collector) latest() float64 {
|
|
c.mtx.RLock()
|
|
defer c.mtx.RUnlock()
|
|
var latest float64
|
|
for iv := range c.expected {
|
|
if iv.end > latest {
|
|
latest = iv.end
|
|
}
|
|
}
|
|
return latest
|
|
}
|
|
|
|
// Want declares that the Collector expects to receive the given alerts
|
|
// within the given time boundaries.
|
|
func (c *Collector) Want(iv Interval, alerts ...*TestAlert) {
|
|
c.mtx.Lock()
|
|
defer c.mtx.Unlock()
|
|
var nas model.Alerts
|
|
for _, a := range alerts {
|
|
nas = append(nas, a.nativeAlert(c.opts))
|
|
}
|
|
|
|
c.expected[iv] = append(c.expected[iv], nas)
|
|
}
|
|
|
|
// add the given alerts to the collected alerts.
|
|
func (c *Collector) add(alerts ...*model.Alert) {
|
|
c.mtx.Lock()
|
|
defer c.mtx.Unlock()
|
|
arrival := c.opts.relativeTime(time.Now())
|
|
|
|
c.collected[arrival] = append(c.collected[arrival], model.Alerts(alerts))
|
|
}
|
|
|
|
func (c *Collector) check() string {
|
|
report := fmt.Sprintf("\ncollector %q:\n\n", c)
|
|
|
|
c.mtx.RLock()
|
|
defer c.mtx.RUnlock()
|
|
for iv, expected := range c.expected {
|
|
report += fmt.Sprintf("interval %v\n", iv)
|
|
|
|
var alerts []model.Alerts
|
|
for at, got := range c.collected {
|
|
if iv.contains(at) {
|
|
alerts = append(alerts, got...)
|
|
}
|
|
}
|
|
|
|
for _, exp := range expected {
|
|
found := len(exp) == 0 && len(alerts) == 0
|
|
|
|
report += fmt.Sprintf("---\n")
|
|
|
|
for _, e := range exp {
|
|
report += fmt.Sprintf("- %v\n", c.opts.alertString(e))
|
|
}
|
|
|
|
for _, a := range alerts {
|
|
if batchesEqual(exp, a, c.opts) {
|
|
found = true
|
|
break
|
|
}
|
|
}
|
|
|
|
if found {
|
|
report += fmt.Sprintf(" [ ✓ ]\n")
|
|
} else {
|
|
c.t.Fail()
|
|
report += fmt.Sprintf(" [ ✗ ]\n")
|
|
}
|
|
}
|
|
}
|
|
|
|
// Detect unexpected notifications.
|
|
var totalExp, totalAct int
|
|
for _, exp := range c.expected {
|
|
for _, e := range exp {
|
|
totalExp += len(e)
|
|
}
|
|
}
|
|
for _, act := range c.collected {
|
|
for _, a := range act {
|
|
if len(a) == 0 {
|
|
c.t.Error("received empty notifications")
|
|
}
|
|
totalAct += len(a)
|
|
}
|
|
}
|
|
if totalExp != totalAct {
|
|
c.t.Fail()
|
|
report += fmt.Sprintf("\nExpected total of %d alerts, got %d", totalExp, totalAct)
|
|
}
|
|
|
|
if c.t.Failed() {
|
|
report += "\nreceived:\n"
|
|
|
|
for at, col := range c.collected {
|
|
for _, alerts := range col {
|
|
report += fmt.Sprintf("@ %v\n", at)
|
|
for _, a := range alerts {
|
|
report += fmt.Sprintf("- %v\n", c.opts.alertString(a))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return report
|
|
}
|