2015-01-21 19:07:45 +00:00
|
|
|
// Copyright 2014 The Prometheus Authors
|
2014-09-19 16:18:44 +00:00
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
2014-09-16 13:47:24 +00:00
|
|
|
package local
|
2014-06-06 09:55:53 +00:00
|
|
|
|
|
|
|
import (
|
2014-09-23 17:21:10 +00:00
|
|
|
"reflect"
|
2014-06-06 09:55:53 +00:00
|
|
|
"testing"
|
|
|
|
|
|
|
|
clientmodel "github.com/prometheus/client_golang/model"
|
|
|
|
|
2014-09-23 17:21:10 +00:00
|
|
|
"github.com/prometheus/prometheus/storage/local/codable"
|
2014-09-14 13:33:56 +00:00
|
|
|
"github.com/prometheus/prometheus/storage/local/index"
|
2014-06-06 09:55:53 +00:00
|
|
|
"github.com/prometheus/prometheus/storage/metric"
|
2014-08-13 15:13:28 +00:00
|
|
|
"github.com/prometheus/prometheus/utility/test"
|
2014-06-06 09:55:53 +00:00
|
|
|
)
|
|
|
|
|
2014-10-28 18:01:41 +00:00
|
|
|
var (
|
|
|
|
m1 = clientmodel.Metric{"label": "value1"}
|
|
|
|
m2 = clientmodel.Metric{"label": "value2"}
|
|
|
|
m3 = clientmodel.Metric{"label": "value3"}
|
|
|
|
)
|
|
|
|
|
2014-10-07 17:11:24 +00:00
|
|
|
func newTestPersistence(t *testing.T) (*persistence, test.Closer) {
|
2014-08-13 15:13:28 +00:00
|
|
|
dir := test.NewTemporaryDirectory("test_persistence", t)
|
2014-11-05 19:02:45 +00:00
|
|
|
p, err := newPersistence(dir.Path(), 1024, false)
|
2014-06-06 09:55:53 +00:00
|
|
|
if err != nil {
|
2014-08-13 15:13:28 +00:00
|
|
|
dir.Close()
|
2014-06-06 09:55:53 +00:00
|
|
|
t.Fatal(err)
|
|
|
|
}
|
2014-08-21 20:06:11 +00:00
|
|
|
return p, test.NewCallbackCloser(func() {
|
2014-10-07 17:11:24 +00:00
|
|
|
p.close()
|
2014-08-21 20:06:11 +00:00
|
|
|
dir.Close()
|
|
|
|
})
|
2014-08-12 15:46:46 +00:00
|
|
|
}
|
2014-08-13 15:13:28 +00:00
|
|
|
|
2014-10-15 13:53:05 +00:00
|
|
|
func buildTestChunks() map[clientmodel.Fingerprint][]chunk {
|
2014-08-13 15:13:28 +00:00
|
|
|
fps := clientmodel.Fingerprints{
|
2014-10-28 18:01:41 +00:00
|
|
|
m1.Fingerprint(),
|
|
|
|
m2.Fingerprint(),
|
|
|
|
m3.Fingerprint(),
|
2014-08-13 15:13:28 +00:00
|
|
|
}
|
2014-10-15 13:53:05 +00:00
|
|
|
fpToChunks := map[clientmodel.Fingerprint][]chunk{}
|
2014-08-13 15:13:28 +00:00
|
|
|
|
|
|
|
for _, fp := range fps {
|
2014-10-15 13:53:05 +00:00
|
|
|
fpToChunks[fp] = make([]chunk, 0, 10)
|
2014-08-13 15:13:28 +00:00
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
fpToChunks[fp] = append(fpToChunks[fp], newDeltaEncodedChunk(d1, d1, true).add(&metric.SamplePair{
|
|
|
|
Timestamp: clientmodel.Timestamp(i),
|
|
|
|
Value: clientmodel.SampleValue(fp),
|
|
|
|
})[0])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return fpToChunks
|
|
|
|
}
|
|
|
|
|
|
|
|
func chunksEqual(c1, c2 chunk) bool {
|
|
|
|
values2 := c2.values()
|
|
|
|
for v1 := range c1.values() {
|
|
|
|
v2 := <-values2
|
|
|
|
if !v1.Equal(v2) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2014-10-28 18:01:41 +00:00
|
|
|
func TestPersistLoadDropChunks(t *testing.T) {
|
2014-08-13 15:13:28 +00:00
|
|
|
p, closer := newTestPersistence(t)
|
|
|
|
defer closer.Close()
|
|
|
|
|
|
|
|
fpToChunks := buildTestChunks()
|
|
|
|
|
|
|
|
for fp, chunks := range fpToChunks {
|
2014-10-27 19:40:48 +00:00
|
|
|
for i, c := range chunks {
|
Improve persisting chunks to disk.
This is done by bucketing chunks by fingerprint. If the persisting to
disk falls behind, more and more chunks are in the queue. As soon as
there are "double hits", we will now persist both chunks in one go,
doubling the disk throughput (assuming it is limited by disk
seeks). Should even more pile up so that we end wit "triple hits", we
will persist those first, and so on.
Even if we have millions of time series, this will still help,
assuming not all of them are growing with the same speed. Series that
get many samples and/or are not very compressable will accumulate
chunks faster, and they will soon get double- or triple-writes.
To improve the chance of double writes,
-storage.local.persistence-queue-capacity could be set to a higher
value. However, that will slow down shutdown a lot (as the queue has
to be worked through). So we leave it to the user to set it to a
really high value. A more fundamental solution would be to checkpoint
not only head chunks, but also chunks still in the persist queue. That
would be quite complicated for a rather limited use-case (running many
time series with high ingestion rate on slow spinning disks).
2015-02-13 19:08:52 +00:00
|
|
|
index, err := p.persistChunks(fp, []chunk{c})
|
2014-10-27 19:40:48 +00:00
|
|
|
if err != nil {
|
2014-08-13 15:13:28 +00:00
|
|
|
t.Fatal(err)
|
|
|
|
}
|
2014-10-27 19:40:48 +00:00
|
|
|
if i != index {
|
|
|
|
t.Errorf("Want chunk index %d, got %d.", i, index)
|
|
|
|
}
|
2014-08-13 15:13:28 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for fp, expectedChunks := range fpToChunks {
|
|
|
|
indexes := make([]int, 0, len(expectedChunks))
|
2014-08-19 16:14:44 +00:00
|
|
|
for i := range expectedChunks {
|
2014-08-13 15:13:28 +00:00
|
|
|
indexes = append(indexes, i)
|
|
|
|
}
|
2014-10-27 19:40:48 +00:00
|
|
|
actualChunks, err := p.loadChunks(fp, indexes, 0)
|
2014-08-13 15:13:28 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
for _, i := range indexes {
|
|
|
|
if !chunksEqual(expectedChunks[i], actualChunks[i]) {
|
2014-10-28 18:01:41 +00:00
|
|
|
t.Errorf("%d. Chunks not equal.", i)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Load all chunk descs.
|
|
|
|
actualChunkDescs, err := p.loadChunkDescs(fp, 10)
|
|
|
|
if len(actualChunkDescs) != 10 {
|
|
|
|
t.Errorf("Got %d chunkDescs, want %d.", len(actualChunkDescs), 10)
|
|
|
|
}
|
|
|
|
for i, cd := range actualChunkDescs {
|
|
|
|
if cd.firstTime() != clientmodel.Timestamp(i) || cd.lastTime() != clientmodel.Timestamp(i) {
|
|
|
|
t.Errorf(
|
|
|
|
"Want ts=%v, got firstTime=%v, lastTime=%v.",
|
|
|
|
i, cd.firstTime(), cd.lastTime(),
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
// Load chunk descs partially.
|
|
|
|
actualChunkDescs, err = p.loadChunkDescs(fp, 5)
|
|
|
|
if len(actualChunkDescs) != 5 {
|
|
|
|
t.Errorf("Got %d chunkDescs, want %d.", len(actualChunkDescs), 5)
|
|
|
|
}
|
|
|
|
for i, cd := range actualChunkDescs {
|
|
|
|
if cd.firstTime() != clientmodel.Timestamp(i) || cd.lastTime() != clientmodel.Timestamp(i) {
|
|
|
|
t.Errorf(
|
|
|
|
"Want ts=%v, got firstTime=%v, lastTime=%v.",
|
|
|
|
i, cd.firstTime(), cd.lastTime(),
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Drop half of the chunks.
|
|
|
|
for fp, expectedChunks := range fpToChunks {
|
2014-11-10 17:22:08 +00:00
|
|
|
firstTime, numDropped, allDropped, err := p.dropChunks(fp, 5)
|
2014-10-28 18:01:41 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
2014-11-10 17:22:08 +00:00
|
|
|
if firstTime != 5 {
|
|
|
|
t.Errorf("want first time 5, got %d", firstTime)
|
|
|
|
}
|
2014-10-28 18:01:41 +00:00
|
|
|
if numDropped != 5 {
|
|
|
|
t.Errorf("want 5 dropped chunks, got %v", numDropped)
|
|
|
|
}
|
|
|
|
if allDropped {
|
|
|
|
t.Error("all chunks dropped")
|
|
|
|
}
|
|
|
|
indexes := make([]int, 5)
|
|
|
|
for i := range indexes {
|
|
|
|
indexes[i] = i
|
|
|
|
}
|
|
|
|
actualChunks, err := p.loadChunks(fp, indexes, 0)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
for _, i := range indexes {
|
|
|
|
if !chunksEqual(expectedChunks[i+5], actualChunks[i]) {
|
|
|
|
t.Errorf("%d. Chunks not equal.", i)
|
2014-08-13 15:13:28 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2014-10-28 18:01:41 +00:00
|
|
|
// Drop all the chunks.
|
|
|
|
for fp := range fpToChunks {
|
2014-11-10 17:22:08 +00:00
|
|
|
firstTime, numDropped, allDropped, err := p.dropChunks(fp, 100)
|
|
|
|
if firstTime != 0 {
|
|
|
|
t.Errorf("want first time 0, got %d", firstTime)
|
|
|
|
}
|
2014-10-28 18:01:41 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if numDropped != 5 {
|
|
|
|
t.Errorf("want 5 dropped chunks, got %v", numDropped)
|
|
|
|
}
|
|
|
|
if !allDropped {
|
|
|
|
t.Error("not all chunks dropped")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestCheckpointAndLoadSeriesMapAndHeads(t *testing.T) {
|
|
|
|
p, closer := newTestPersistence(t)
|
|
|
|
defer closer.Close()
|
|
|
|
|
|
|
|
fpLocker := newFingerprintLocker(10)
|
|
|
|
sm := newSeriesMap()
|
2014-11-05 19:02:45 +00:00
|
|
|
s1 := newMemorySeries(m1, true, 0)
|
|
|
|
s2 := newMemorySeries(m2, false, 0)
|
|
|
|
s3 := newMemorySeries(m3, false, 0)
|
2014-10-28 18:01:41 +00:00
|
|
|
s1.add(m1.Fingerprint(), &metric.SamplePair{Timestamp: 1, Value: 3.14})
|
|
|
|
s3.add(m1.Fingerprint(), &metric.SamplePair{Timestamp: 2, Value: 2.7})
|
|
|
|
s3.headChunkPersisted = true
|
|
|
|
sm.put(m1.Fingerprint(), s1)
|
|
|
|
sm.put(m2.Fingerprint(), s2)
|
|
|
|
sm.put(m3.Fingerprint(), s3)
|
|
|
|
|
|
|
|
if err := p.checkpointSeriesMapAndHeads(sm, fpLocker); err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
loadedSM, err := p.loadSeriesMapAndHeads()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if loadedSM.length() != 2 {
|
|
|
|
t.Errorf("want 2 series in map, got %d", loadedSM.length())
|
|
|
|
}
|
|
|
|
if loadedS1, ok := loadedSM.get(m1.Fingerprint()); ok {
|
|
|
|
if !reflect.DeepEqual(loadedS1.metric, m1) {
|
|
|
|
t.Errorf("want metric %v, got %v", m1, loadedS1.metric)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(loadedS1.head().chunk, s1.head().chunk) {
|
|
|
|
t.Error("head chunks differ")
|
|
|
|
}
|
|
|
|
if loadedS1.chunkDescsOffset != 0 {
|
|
|
|
t.Errorf("want chunkDescsOffset 0, got %d", loadedS1.chunkDescsOffset)
|
|
|
|
}
|
|
|
|
if loadedS1.headChunkPersisted {
|
|
|
|
t.Error("headChunkPersisted is true")
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
t.Errorf("couldn't find %v in loaded map", m1)
|
|
|
|
}
|
|
|
|
if loadedS3, ok := loadedSM.get(m3.Fingerprint()); ok {
|
|
|
|
if !reflect.DeepEqual(loadedS3.metric, m3) {
|
|
|
|
t.Errorf("want metric %v, got %v", m3, loadedS3.metric)
|
|
|
|
}
|
|
|
|
if loadedS3.head().chunk != nil {
|
|
|
|
t.Error("head chunk not evicted")
|
|
|
|
}
|
|
|
|
if loadedS3.chunkDescsOffset != -1 {
|
|
|
|
t.Errorf("want chunkDescsOffset -1, got %d", loadedS3.chunkDescsOffset)
|
|
|
|
}
|
|
|
|
if !loadedS3.headChunkPersisted {
|
|
|
|
t.Error("headChunkPersisted is false")
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
t.Errorf("couldn't find %v in loaded map", m1)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestGetFingerprintsModifiedBefore(t *testing.T) {
|
|
|
|
p, closer := newTestPersistence(t)
|
|
|
|
defer closer.Close()
|
|
|
|
|
|
|
|
m1 := clientmodel.Metric{"n1": "v1"}
|
|
|
|
m2 := clientmodel.Metric{"n2": "v2"}
|
|
|
|
m3 := clientmodel.Metric{"n1": "v2"}
|
|
|
|
p.archiveMetric(1, m1, 2, 4)
|
|
|
|
p.archiveMetric(2, m2, 1, 6)
|
|
|
|
p.archiveMetric(3, m3, 5, 5)
|
|
|
|
|
|
|
|
expectedFPs := map[clientmodel.Timestamp][]clientmodel.Fingerprint{
|
|
|
|
0: {},
|
|
|
|
1: {},
|
|
|
|
2: {2},
|
|
|
|
3: {1, 2},
|
|
|
|
4: {1, 2},
|
|
|
|
5: {1, 2},
|
|
|
|
6: {1, 2, 3},
|
|
|
|
}
|
|
|
|
|
|
|
|
for ts, want := range expectedFPs {
|
|
|
|
got, err := p.getFingerprintsModifiedBefore(ts)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(want, got) {
|
|
|
|
t.Errorf("timestamp: %v, want FPs %v, got %v", ts, want, got)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-11-05 19:02:45 +00:00
|
|
|
unarchived, firstTime, err := p.unarchiveMetric(1)
|
2014-10-28 18:01:41 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !unarchived {
|
|
|
|
t.Fatal("expected actual unarchival")
|
|
|
|
}
|
2014-11-05 19:02:45 +00:00
|
|
|
if firstTime != 2 {
|
|
|
|
t.Errorf("expected first time 2, got %v", firstTime)
|
|
|
|
}
|
|
|
|
unarchived, firstTime, err = p.unarchiveMetric(1)
|
2014-10-28 18:01:41 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if unarchived {
|
|
|
|
t.Fatal("expected no unarchival")
|
|
|
|
}
|
|
|
|
|
|
|
|
expectedFPs = map[clientmodel.Timestamp][]clientmodel.Fingerprint{
|
|
|
|
0: {},
|
|
|
|
1: {},
|
|
|
|
2: {2},
|
|
|
|
3: {2},
|
|
|
|
4: {2},
|
|
|
|
5: {2},
|
|
|
|
6: {2, 3},
|
|
|
|
}
|
|
|
|
|
|
|
|
for ts, want := range expectedFPs {
|
|
|
|
got, err := p.getFingerprintsModifiedBefore(ts)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !reflect.DeepEqual(want, got) {
|
|
|
|
t.Errorf("timestamp: %v, want FPs %v, got %v", ts, want, got)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestDropArchivedMetric(t *testing.T) {
|
|
|
|
p, closer := newTestPersistence(t)
|
|
|
|
defer closer.Close()
|
|
|
|
|
|
|
|
m1 := clientmodel.Metric{"n1": "v1"}
|
|
|
|
m2 := clientmodel.Metric{"n2": "v2"}
|
|
|
|
p.archiveMetric(1, m1, 2, 4)
|
|
|
|
p.archiveMetric(2, m2, 1, 6)
|
|
|
|
p.indexMetric(1, m1)
|
|
|
|
p.indexMetric(2, m2)
|
|
|
|
p.waitForIndexing()
|
|
|
|
|
|
|
|
outFPs, err := p.getFingerprintsForLabelPair(metric.LabelPair{Name: "n1", Value: "v1"})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
want := clientmodel.Fingerprints{1}
|
|
|
|
if !reflect.DeepEqual(outFPs, want) {
|
|
|
|
t.Errorf("want %#v, got %#v", want, outFPs)
|
|
|
|
}
|
|
|
|
outFPs, err = p.getFingerprintsForLabelPair(metric.LabelPair{Name: "n2", Value: "v2"})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
want = clientmodel.Fingerprints{2}
|
|
|
|
if !reflect.DeepEqual(outFPs, want) {
|
|
|
|
t.Errorf("want %#v, got %#v", want, outFPs)
|
|
|
|
}
|
|
|
|
if archived, _, _, err := p.hasArchivedMetric(1); err != nil || !archived {
|
|
|
|
t.Error("want FP 1 archived")
|
|
|
|
}
|
|
|
|
if archived, _, _, err := p.hasArchivedMetric(2); err != nil || !archived {
|
|
|
|
t.Error("want FP 2 archived")
|
|
|
|
}
|
|
|
|
|
2015-02-26 14:19:44 +00:00
|
|
|
if err != p.purgeArchivedMetric(1) {
|
2014-10-28 18:01:41 +00:00
|
|
|
t.Fatal(err)
|
|
|
|
}
|
2015-02-26 14:19:44 +00:00
|
|
|
if err != p.purgeArchivedMetric(3) {
|
|
|
|
// Purging something that has not beet archived is not an error.
|
2014-10-28 18:01:41 +00:00
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
p.waitForIndexing()
|
|
|
|
|
|
|
|
outFPs, err = p.getFingerprintsForLabelPair(metric.LabelPair{Name: "n1", Value: "v1"})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
want = nil
|
|
|
|
if !reflect.DeepEqual(outFPs, want) {
|
|
|
|
t.Errorf("want %#v, got %#v", want, outFPs)
|
|
|
|
}
|
|
|
|
outFPs, err = p.getFingerprintsForLabelPair(metric.LabelPair{Name: "n2", Value: "v2"})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
want = clientmodel.Fingerprints{2}
|
|
|
|
if !reflect.DeepEqual(outFPs, want) {
|
|
|
|
t.Errorf("want %#v, got %#v", want, outFPs)
|
|
|
|
}
|
|
|
|
if archived, _, _, err := p.hasArchivedMetric(1); err != nil || archived {
|
|
|
|
t.Error("want FP 1 not archived")
|
|
|
|
}
|
|
|
|
if archived, _, _, err := p.hasArchivedMetric(2); err != nil || !archived {
|
|
|
|
t.Error("want FP 2 archived")
|
|
|
|
}
|
2014-08-13 15:13:28 +00:00
|
|
|
}
|
2014-09-14 13:33:56 +00:00
|
|
|
|
|
|
|
type incrementalBatch struct {
|
|
|
|
fpToMetric index.FingerprintMetricMapping
|
|
|
|
expectedLnToLvs index.LabelNameLabelValuesMapping
|
|
|
|
expectedLpToFps index.LabelPairFingerprintsMapping
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestIndexing(t *testing.T) {
|
|
|
|
batches := []incrementalBatch{
|
|
|
|
{
|
|
|
|
fpToMetric: index.FingerprintMetricMapping{
|
|
|
|
0: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_0",
|
|
|
|
"label_1": "value_1",
|
|
|
|
},
|
|
|
|
1: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_0",
|
|
|
|
"label_2": "value_2",
|
|
|
|
"label_3": "value_3",
|
|
|
|
},
|
|
|
|
2: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_1",
|
|
|
|
"label_1": "value_2",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
expectedLnToLvs: index.LabelNameLabelValuesMapping{
|
2014-09-23 17:21:10 +00:00
|
|
|
clientmodel.MetricNameLabel: codable.LabelValueSet{
|
|
|
|
"metric_0": struct{}{},
|
|
|
|
"metric_1": struct{}{},
|
|
|
|
},
|
|
|
|
"label_1": codable.LabelValueSet{
|
|
|
|
"value_1": struct{}{},
|
|
|
|
"value_2": struct{}{},
|
|
|
|
},
|
|
|
|
"label_2": codable.LabelValueSet{
|
|
|
|
"value_2": struct{}{},
|
|
|
|
},
|
|
|
|
"label_3": codable.LabelValueSet{
|
|
|
|
"value_3": struct{}{},
|
|
|
|
},
|
2014-09-14 13:33:56 +00:00
|
|
|
},
|
|
|
|
expectedLpToFps: index.LabelPairFingerprintsMapping{
|
|
|
|
metric.LabelPair{
|
|
|
|
Name: clientmodel.MetricNameLabel,
|
|
|
|
Value: "metric_0",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{0: struct{}{}, 1: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: clientmodel.MetricNameLabel,
|
|
|
|
Value: "metric_1",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{2: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_1",
|
|
|
|
Value: "value_1",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{0: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_1",
|
|
|
|
Value: "value_2",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{2: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_2",
|
|
|
|
Value: "value_2",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{1: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_3",
|
|
|
|
Value: "value_3",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{1: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
},
|
|
|
|
}, {
|
|
|
|
fpToMetric: index.FingerprintMetricMapping{
|
|
|
|
3: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_0",
|
|
|
|
"label_1": "value_3",
|
|
|
|
},
|
|
|
|
4: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_2",
|
|
|
|
"label_2": "value_2",
|
|
|
|
"label_3": "value_1",
|
|
|
|
},
|
|
|
|
5: {
|
|
|
|
clientmodel.MetricNameLabel: "metric_1",
|
|
|
|
"label_1": "value_3",
|
|
|
|
},
|
|
|
|
},
|
|
|
|
expectedLnToLvs: index.LabelNameLabelValuesMapping{
|
2014-09-23 17:21:10 +00:00
|
|
|
clientmodel.MetricNameLabel: codable.LabelValueSet{
|
|
|
|
"metric_0": struct{}{},
|
|
|
|
"metric_1": struct{}{},
|
|
|
|
"metric_2": struct{}{},
|
|
|
|
},
|
|
|
|
"label_1": codable.LabelValueSet{
|
|
|
|
"value_1": struct{}{},
|
|
|
|
"value_2": struct{}{},
|
|
|
|
"value_3": struct{}{},
|
|
|
|
},
|
|
|
|
"label_2": codable.LabelValueSet{
|
|
|
|
"value_2": struct{}{},
|
|
|
|
},
|
|
|
|
"label_3": codable.LabelValueSet{
|
|
|
|
"value_1": struct{}{},
|
|
|
|
"value_3": struct{}{},
|
|
|
|
},
|
2014-09-14 13:33:56 +00:00
|
|
|
},
|
|
|
|
expectedLpToFps: index.LabelPairFingerprintsMapping{
|
|
|
|
metric.LabelPair{
|
|
|
|
Name: clientmodel.MetricNameLabel,
|
|
|
|
Value: "metric_0",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{0: struct{}{}, 1: struct{}{}, 3: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: clientmodel.MetricNameLabel,
|
|
|
|
Value: "metric_1",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{2: struct{}{}, 5: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: clientmodel.MetricNameLabel,
|
|
|
|
Value: "metric_2",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{4: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_1",
|
|
|
|
Value: "value_1",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{0: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_1",
|
|
|
|
Value: "value_2",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{2: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_1",
|
|
|
|
Value: "value_3",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{3: struct{}{}, 5: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_2",
|
|
|
|
Value: "value_2",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{1: struct{}{}, 4: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_3",
|
|
|
|
Value: "value_1",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{4: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
metric.LabelPair{
|
|
|
|
Name: "label_3",
|
|
|
|
Value: "value_3",
|
2014-09-23 17:21:10 +00:00
|
|
|
}: codable.FingerprintSet{1: struct{}{}},
|
2014-09-14 13:33:56 +00:00
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
p, closer := newTestPersistence(t)
|
|
|
|
defer closer.Close()
|
|
|
|
|
|
|
|
indexedFpsToMetrics := index.FingerprintMetricMapping{}
|
|
|
|
for i, b := range batches {
|
|
|
|
for fp, m := range b.fpToMetric {
|
2014-10-28 18:01:41 +00:00
|
|
|
p.indexMetric(fp, m)
|
2014-10-07 17:11:24 +00:00
|
|
|
if err := p.archiveMetric(fp, m, 1, 2); err != nil {
|
2014-09-14 13:33:56 +00:00
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
indexedFpsToMetrics[fp] = m
|
|
|
|
}
|
2014-10-07 17:11:24 +00:00
|
|
|
verifyIndexedState(i, t, b, indexedFpsToMetrics, p)
|
2014-09-14 13:33:56 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
for i := len(batches) - 1; i >= 0; i-- {
|
|
|
|
b := batches[i]
|
2014-10-07 17:11:24 +00:00
|
|
|
verifyIndexedState(i, t, batches[i], indexedFpsToMetrics, p)
|
2014-09-14 13:33:56 +00:00
|
|
|
for fp, m := range b.fpToMetric {
|
2014-10-28 18:01:41 +00:00
|
|
|
p.unindexMetric(fp, m)
|
2014-11-05 19:02:45 +00:00
|
|
|
unarchived, firstTime, err := p.unarchiveMetric(fp)
|
2014-09-14 13:33:56 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !unarchived {
|
|
|
|
t.Errorf("%d. metric not unarchived", i)
|
|
|
|
}
|
2014-11-05 19:02:45 +00:00
|
|
|
if firstTime != 1 {
|
|
|
|
t.Errorf("%d. expected firstTime=1, got %v", i, firstTime)
|
|
|
|
}
|
2014-09-14 13:33:56 +00:00
|
|
|
delete(indexedFpsToMetrics, fp)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-10-07 17:11:24 +00:00
|
|
|
func verifyIndexedState(i int, t *testing.T, b incrementalBatch, indexedFpsToMetrics index.FingerprintMetricMapping, p *persistence) {
|
|
|
|
p.waitForIndexing()
|
2014-09-14 13:33:56 +00:00
|
|
|
for fp, m := range indexedFpsToMetrics {
|
|
|
|
// Compare archived metrics with input metrics.
|
2014-10-07 17:11:24 +00:00
|
|
|
mOut, err := p.getArchivedMetric(fp)
|
2014-09-14 13:33:56 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !mOut.Equal(m) {
|
|
|
|
t.Errorf("%d. %v: Got: %s; want %s", i, fp, mOut, m)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check that archived metrics are in membership index.
|
2014-10-07 17:11:24 +00:00
|
|
|
has, first, last, err := p.hasArchivedMetric(fp)
|
2014-09-14 13:33:56 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Errorf("%d. fingerprint %v not found", i, fp)
|
|
|
|
}
|
|
|
|
if first != 1 || last != 2 {
|
|
|
|
t.Errorf(
|
|
|
|
"%d. %v: Got first: %d, last %d; want first: %d, last %d",
|
|
|
|
i, fp, first, last, 1, 2,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compare label name -> label values mappings.
|
|
|
|
for ln, lvs := range b.expectedLnToLvs {
|
2014-10-07 17:11:24 +00:00
|
|
|
outLvs, err := p.getLabelValuesForLabelName(ln)
|
2014-09-14 13:33:56 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
|
2014-09-23 17:21:10 +00:00
|
|
|
outSet := codable.LabelValueSet{}
|
|
|
|
for _, lv := range outLvs {
|
|
|
|
outSet[lv] = struct{}{}
|
2014-09-14 13:33:56 +00:00
|
|
|
}
|
2014-09-23 17:21:10 +00:00
|
|
|
|
|
|
|
if !reflect.DeepEqual(lvs, outSet) {
|
|
|
|
t.Errorf("%d. label values don't match. Got: %v; want %v", i, outSet, lvs)
|
2014-09-14 13:33:56 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compare label pair -> fingerprints mappings.
|
|
|
|
for lp, fps := range b.expectedLpToFps {
|
2014-10-28 18:01:41 +00:00
|
|
|
outFPs, err := p.getFingerprintsForLabelPair(lp)
|
2014-09-14 13:33:56 +00:00
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
|
2014-09-23 17:21:10 +00:00
|
|
|
outSet := codable.FingerprintSet{}
|
2014-10-28 18:01:41 +00:00
|
|
|
for _, fp := range outFPs {
|
2014-09-23 17:21:10 +00:00
|
|
|
outSet[fp] = struct{}{}
|
2014-09-14 13:33:56 +00:00
|
|
|
}
|
2014-09-23 17:21:10 +00:00
|
|
|
|
|
|
|
if !reflect.DeepEqual(fps, outSet) {
|
|
|
|
t.Errorf("%d. %v: fingerprints don't match. Got: %v; want %v", i, lp, outSet, fps)
|
2014-09-14 13:33:56 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|