// Copyright 2017 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package tsdb import ( "context" "fmt" "io/ioutil" "math" "os" "path" "path/filepath" "testing" "time" "github.com/go-kit/kit/log" "github.com/pkg/errors" prom_testutil "github.com/prometheus/client_golang/prometheus/testutil" dto "github.com/prometheus/client_model/go" "github.com/prometheus/tsdb/chunks" "github.com/prometheus/tsdb/fileutil" "github.com/prometheus/tsdb/labels" "github.com/prometheus/tsdb/testutil" ) func TestSplitByRange(t *testing.T) { cases := []struct { trange int64 ranges [][2]int64 output [][][2]int64 }{ { trange: 60, ranges: [][2]int64{{0, 10}}, output: [][][2]int64{ {{0, 10}}, }, }, { trange: 60, ranges: [][2]int64{{0, 60}}, output: [][][2]int64{ {{0, 60}}, }, }, { trange: 60, ranges: [][2]int64{{0, 10}, {9, 15}, {30, 60}}, output: [][][2]int64{ {{0, 10}, {9, 15}, {30, 60}}, }, }, { trange: 60, ranges: [][2]int64{{70, 90}, {125, 130}, {130, 180}, {1000, 1001}}, output: [][][2]int64{ {{70, 90}}, {{125, 130}, {130, 180}}, {{1000, 1001}}, }, }, // Mis-aligned or too-large blocks are ignored. { trange: 60, ranges: [][2]int64{{50, 70}, {70, 80}}, output: [][][2]int64{ {{70, 80}}, }, }, { trange: 72, ranges: [][2]int64{{0, 144}, {144, 216}, {216, 288}}, output: [][][2]int64{ {{144, 216}}, {{216, 288}}, }, }, // Various awkward edge cases easy to hit with negative numbers. { trange: 60, ranges: [][2]int64{{-10, -5}}, output: [][][2]int64{ {{-10, -5}}, }, }, { trange: 60, ranges: [][2]int64{{-60, -50}, {-10, -5}}, output: [][][2]int64{ {{-60, -50}, {-10, -5}}, }, }, { trange: 60, ranges: [][2]int64{{-60, -50}, {-10, -5}, {0, 15}}, output: [][][2]int64{ {{-60, -50}, {-10, -5}}, {{0, 15}}, }, }, } for _, c := range cases { // Transform input range tuples into dirMetas. blocks := make([]dirMeta, 0, len(c.ranges)) for _, r := range c.ranges { blocks = append(blocks, dirMeta{ meta: &BlockMeta{ MinTime: r[0], MaxTime: r[1], }, }) } // Transform output range tuples into dirMetas. exp := make([][]dirMeta, len(c.output)) for i, group := range c.output { for _, r := range group { exp[i] = append(exp[i], dirMeta{ meta: &BlockMeta{MinTime: r[0], MaxTime: r[1]}, }) } } testutil.Equals(t, exp, splitByRange(blocks, c.trange)) } } // See https://github.com/prometheus/prometheus/issues/3064 func TestNoPanicFor0Tombstones(t *testing.T) { metas := []dirMeta{ { dir: "1", meta: &BlockMeta{ MinTime: 0, MaxTime: 100, }, }, { dir: "2", meta: &BlockMeta{ MinTime: 101, MaxTime: 200, }, }, } c, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{50}, nil) testutil.Ok(t, err) c.plan(metas) } func TestLeveledCompactor_plan(t *testing.T) { // This mimicks our default ExponentialBlockRanges with min block size equals to 20. compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{ 20, 60, 180, 540, 1620, }, nil) testutil.Ok(t, err) cases := []struct { metas []dirMeta expected []string }{ { metas: []dirMeta{ metaRange("1", 0, 20, nil), }, expected: nil, }, // We should wait for four blocks of size 20 to appear before compacting. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), }, expected: nil, }, // We should wait for a next block of size 20 to appear before compacting // the existing ones. We have three, but we ignore the fresh one from WAl. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 40, 60, nil), }, expected: nil, }, // Block to fill the entire parent range appeared – should be compacted. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 40, 60, nil), metaRange("4", 60, 80, nil), }, expected: []string{"1", "2", "3"}, }, // Block for the next parent range appeared with gap with size 20. Nothing will happen in the first one // anymore but we ignore fresh one still, so no compaction. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 60, 80, nil), }, expected: nil, }, // Block for the next parent range appeared, and we have a gap with size 20 between second and third block. // We will not get this missed gap anymore and we should compact just these two. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 60, 80, nil), metaRange("4", 80, 100, nil), }, expected: []string{"1", "2"}, }, { // We have 20, 20, 20, 60, 60 range blocks. "5" is marked as fresh one. metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 40, 60, nil), metaRange("4", 60, 120, nil), metaRange("5", 120, 180, nil), }, expected: []string{"1", "2", "3"}, }, { // We have 20, 60, 20, 60, 240 range blocks. We can compact 20 + 60 + 60. metas: []dirMeta{ metaRange("2", 20, 40, nil), metaRange("4", 60, 120, nil), metaRange("5", 960, 980, nil), // Fresh one. metaRange("6", 120, 180, nil), metaRange("7", 720, 960, nil), }, expected: []string{"2", "4", "6"}, }, // Do not select large blocks that have many tombstones when there is no fresh block. { metas: []dirMeta{ metaRange("1", 0, 540, &BlockStats{ NumSeries: 10, NumTombstones: 3, }), }, expected: nil, }, // Select large blocks that have many tombstones when fresh appears. { metas: []dirMeta{ metaRange("1", 0, 540, &BlockStats{ NumSeries: 10, NumTombstones: 3, }), metaRange("2", 540, 560, nil), }, expected: []string{"1"}, }, // For small blocks, do not compact tombstones, even when fresh appears. { metas: []dirMeta{ metaRange("1", 0, 60, &BlockStats{ NumSeries: 10, NumTombstones: 3, }), metaRange("2", 60, 80, nil), }, expected: nil, }, // Regression test: we were stuck in a compact loop where we always recompacted // the same block when tombstones and series counts were zero. { metas: []dirMeta{ metaRange("1", 0, 540, &BlockStats{ NumSeries: 0, NumTombstones: 0, }), metaRange("2", 540, 560, nil), }, expected: nil, }, // Regression test: we were wrongly assuming that new block is fresh from WAL when its ULID is newest. // We need to actually look on max time instead. // // With previous, wrong approach "8" block was ignored, so we were wrongly compacting 5 and 7 and introducing // block overlaps. { metas: []dirMeta{ metaRange("5", 0, 360, nil), metaRange("6", 540, 560, nil), // Fresh one. metaRange("7", 360, 420, nil), metaRange("8", 420, 540, nil), }, expected: []string{"7", "8"}, }, // For overlapping blocks. { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 19, 40, nil), metaRange("3", 40, 60, nil), }, expected: []string{"1", "2"}, }, { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 30, 50, nil), }, expected: []string{"2", "3"}, }, { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 10, 40, nil), metaRange("3", 30, 50, nil), }, expected: []string{"1", "2", "3"}, }, { metas: []dirMeta{ metaRange("5", 0, 360, nil), metaRange("6", 340, 560, nil), metaRange("7", 360, 420, nil), metaRange("8", 420, 540, nil), }, expected: []string{"5", "6", "7", "8"}, }, } for _, c := range cases { if !t.Run("", func(t *testing.T) { res, err := compactor.plan(c.metas) testutil.Ok(t, err) testutil.Equals(t, c.expected, res) }) { return } } } func TestRangeWithFailedCompactionWontGetSelected(t *testing.T) { compactor, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{ 20, 60, 240, 720, 2160, }, nil) testutil.Ok(t, err) cases := []struct { metas []dirMeta }{ { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 40, 60, nil), metaRange("4", 60, 80, nil), }, }, { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 60, 80, nil), metaRange("4", 80, 100, nil), }, }, { metas: []dirMeta{ metaRange("1", 0, 20, nil), metaRange("2", 20, 40, nil), metaRange("3", 40, 60, nil), metaRange("4", 60, 120, nil), metaRange("5", 120, 180, nil), metaRange("6", 180, 200, nil), }, }, } for _, c := range cases { c.metas[1].meta.Compaction.Failed = true res, err := compactor.plan(c.metas) testutil.Ok(t, err) testutil.Equals(t, []string(nil), res) } } func TestCompactionFailWillCleanUpTempDir(t *testing.T) { compactor, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{ 20, 60, 240, 720, 2160, }, nil) testutil.Ok(t, err) tmpdir, err := ioutil.TempDir("", "test") testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(tmpdir)) }() testutil.NotOk(t, compactor.write(tmpdir, &BlockMeta{}, erringBReader{})) _, err = os.Stat(filepath.Join(tmpdir, BlockMeta{}.ULID.String()) + ".tmp") testutil.Assert(t, os.IsNotExist(err), "directory is not cleaned up") } func metaRange(name string, mint, maxt int64, stats *BlockStats) dirMeta { meta := &BlockMeta{MinTime: mint, MaxTime: maxt} if stats != nil { meta.Stats = *stats } return dirMeta{ dir: name, meta: meta, } } type erringBReader struct{} func (erringBReader) Index() (IndexReader, error) { return nil, errors.New("index") } func (erringBReader) Chunks() (ChunkReader, error) { return nil, errors.New("chunks") } func (erringBReader) Tombstones() (TombstoneReader, error) { return nil, errors.New("tombstones") } func (erringBReader) MinTime() int64 { return 0 } func (erringBReader) MaxTime() int64 { return 0 } type nopChunkWriter struct{} func (nopChunkWriter) WriteChunks(chunks ...chunks.Meta) error { return nil } func (nopChunkWriter) Close() error { return nil } func TestCompaction_populateBlock(t *testing.T) { var populateBlocksCases = []struct { title string inputSeriesSamples [][]seriesSamples compactMinTime int64 compactMaxTime int64 // When not defined the test runner sets a default of math.MaxInt64. expSeriesSamples []seriesSamples expErr error }{ { title: "Populate block from empty input should return error.", inputSeriesSamples: [][]seriesSamples{}, expErr: errors.New("cannot populate block from no readers"), }, { // Populate from single block without chunks. We expect these kind of series being ignored. inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, }, }, }, }, { title: "Populate from single block. We expect the same samples at the output.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}}, }, }, }, { title: "Populate from two blocks.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}}, }, { // no-chunk series should be dropped. lset: map[string]string{"a": "empty"}, }, }, { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 21}, {t: 30}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 40}, {t: 45}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}, {{t: 21}, {t: 30}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}, {{t: 40}, {t: 45}}}, }, }, }, { title: "Populate from two blocks showing that order is maintained.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}}, }, }, { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 21}, {t: 30}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 40}, {t: 45}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 0}, {t: 10}}, {{t: 11}, {t: 20}}, {{t: 21}, {t: 30}}}, }, { lset: map[string]string{"a": "c"}, chunks: [][]sample{{{t: 1}, {t: 9}}, {{t: 10}, {t: 19}}, {{t: 40}, {t: 45}}}, }, }, }, { title: "Populate from two blocks showing that order of series is sorted.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "4"}, chunks: [][]sample{{{t: 5}, {t: 7}}}, }, { lset: map[string]string{"a": "3"}, chunks: [][]sample{{{t: 5}, {t: 6}}}, }, { lset: map[string]string{"a": "same"}, chunks: [][]sample{{{t: 1}, {t: 4}}}, }, }, { { lset: map[string]string{"a": "2"}, chunks: [][]sample{{{t: 1}, {t: 3}}}, }, { lset: map[string]string{"a": "1"}, chunks: [][]sample{{{t: 1}, {t: 2}}}, }, { lset: map[string]string{"a": "same"}, chunks: [][]sample{{{t: 5}, {t: 8}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"a": "1"}, chunks: [][]sample{{{t: 1}, {t: 2}}}, }, { lset: map[string]string{"a": "2"}, chunks: [][]sample{{{t: 1}, {t: 3}}}, }, { lset: map[string]string{"a": "3"}, chunks: [][]sample{{{t: 5}, {t: 6}}}, }, { lset: map[string]string{"a": "4"}, chunks: [][]sample{{{t: 5}, {t: 7}}}, }, { lset: map[string]string{"a": "same"}, chunks: [][]sample{{{t: 1}, {t: 4}}, {{t: 5}, {t: 8}}}, }, }, }, { // This should not happened because head block is making sure the chunks are not crossing block boundaries. title: "Populate from single block containing chunk outside of compact meta time range.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 30}}}, }, }, }, compactMinTime: 0, compactMaxTime: 20, expErr: errors.New("found chunk with minTime: 10 maxTime: 30 outside of compacted minTime: 0 maxTime: 20"), }, { // Introduced by https://github.com/prometheus/tsdb/issues/347. title: "Populate from single block containing extra chunk", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "issue347"}, chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}}, }, }, }, compactMinTime: 0, compactMaxTime: 10, expErr: errors.New("found chunk with minTime: 10 maxTime: 20 outside of compacted minTime: 0 maxTime: 10"), }, { // Deduplication expected. // Introduced by pull/370 and pull/539. title: "Populate from two blocks containing duplicated chunk.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}}, }, }, { { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 10}, {t: 20}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"a": "b"}, chunks: [][]sample{{{t: 1}, {t: 2}}, {{t: 10}, {t: 20}}}, }, }, }, { // Introduced by https://github.com/prometheus/tsdb/pull/539. title: "Populate from three blocks that the last two are overlapping.", inputSeriesSamples: [][]seriesSamples{ { { lset: map[string]string{"before": "fix"}, chunks: [][]sample{{{t: 0}, {t: 10}, {t: 11}, {t: 20}}}, }, { lset: map[string]string{"after": "fix"}, chunks: [][]sample{{{t: 0}, {t: 10}, {t: 11}, {t: 20}}}, }, }, { { lset: map[string]string{"before": "fix"}, chunks: [][]sample{{{t: 19}, {t: 30}}}, }, { lset: map[string]string{"after": "fix"}, chunks: [][]sample{{{t: 21}, {t: 30}}}, }, }, { { lset: map[string]string{"before": "fix"}, chunks: [][]sample{{{t: 27}, {t: 35}}}, }, { lset: map[string]string{"after": "fix"}, chunks: [][]sample{{{t: 27}, {t: 35}}}, }, }, }, expSeriesSamples: []seriesSamples{ { lset: map[string]string{"after": "fix"}, chunks: [][]sample{{{t: 0}, {t: 10}, {t: 11}, {t: 20}}, {{t: 21}, {t: 27}, {t: 30}, {t: 35}}}, }, { lset: map[string]string{"before": "fix"}, chunks: [][]sample{{{t: 0}, {t: 10}, {t: 11}, {t: 19}, {t: 20}, {t: 27}, {t: 30}, {t: 35}}}, }, }, }, } for _, tc := range populateBlocksCases { if ok := t.Run(tc.title, func(t *testing.T) { blocks := make([]BlockReader, 0, len(tc.inputSeriesSamples)) for _, b := range tc.inputSeriesSamples { ir, cr, mint, maxt := createIdxChkReaders(b) blocks = append(blocks, &mockBReader{ir: ir, cr: cr, mint: mint, maxt: maxt}) } c, err := NewLeveledCompactor(context.Background(), nil, nil, []int64{0}, nil) testutil.Ok(t, err) meta := &BlockMeta{ MinTime: tc.compactMinTime, MaxTime: tc.compactMaxTime, } if meta.MaxTime == 0 { meta.MaxTime = math.MaxInt64 } iw := &mockIndexWriter{} err = c.populateBlock(blocks, meta, iw, nopChunkWriter{}) if tc.expErr != nil { testutil.NotOk(t, err) testutil.Equals(t, tc.expErr.Error(), err.Error()) return } testutil.Ok(t, err) testutil.Equals(t, tc.expSeriesSamples, iw.series) // Check if stats are calculated properly. s := BlockStats{ NumSeries: uint64(len(tc.expSeriesSamples)), } for _, series := range tc.expSeriesSamples { s.NumChunks += uint64(len(series.chunks)) for _, chk := range series.chunks { s.NumSamples += uint64(len(chk)) } } testutil.Equals(t, s, meta.Stats) }); !ok { return } } } func BenchmarkCompaction(b *testing.B) { cases := []struct { ranges [][2]int64 compactionType string }{ { ranges: [][2]int64{{0, 100}, {200, 300}, {400, 500}, {600, 700}}, compactionType: "normal", }, { ranges: [][2]int64{{0, 1000}, {2000, 3000}, {4000, 5000}, {6000, 7000}}, compactionType: "normal", }, { ranges: [][2]int64{{0, 10000}, {20000, 30000}, {40000, 50000}, {60000, 70000}}, compactionType: "normal", }, { ranges: [][2]int64{{0, 100000}, {200000, 300000}, {400000, 500000}, {600000, 700000}}, compactionType: "normal", }, // 40% overlaps. { ranges: [][2]int64{{0, 100}, {60, 160}, {120, 220}, {180, 280}}, compactionType: "vertical", }, { ranges: [][2]int64{{0, 1000}, {600, 1600}, {1200, 2200}, {1800, 2800}}, compactionType: "vertical", }, { ranges: [][2]int64{{0, 10000}, {6000, 16000}, {12000, 22000}, {18000, 28000}}, compactionType: "vertical", }, { ranges: [][2]int64{{0, 100000}, {60000, 160000}, {120000, 220000}, {180000, 280000}}, compactionType: "vertical", }, } nSeries := 10000 for _, c := range cases { nBlocks := len(c.ranges) b.Run(fmt.Sprintf("type=%s,blocks=%d,series=%d,samplesPerSeriesPerBlock=%d", c.compactionType, nBlocks, nSeries, c.ranges[0][1]-c.ranges[0][0]+1), func(b *testing.B) { dir, err := ioutil.TempDir("", "bench_compaction") testutil.Ok(b, err) defer func() { testutil.Ok(b, os.RemoveAll(dir)) }() blockDirs := make([]string, 0, len(c.ranges)) var blocks []*Block for _, r := range c.ranges { block, err := OpenBlock(nil, createBlock(b, dir, genSeries(nSeries, 10, r[0], r[1])), nil) testutil.Ok(b, err) blocks = append(blocks, block) defer func() { testutil.Ok(b, block.Close()) }() blockDirs = append(blockDirs, block.Dir()) } c, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{0}, nil) testutil.Ok(b, err) b.ResetTimer() b.ReportAllocs() _, err = c.Compact(dir, blockDirs, blocks) testutil.Ok(b, err) }) } } // TestDisableAutoCompactions checks that we can // disable and enable the auto compaction. // This is needed for unit tests that rely on // checking state before and after a compaction. func TestDisableAutoCompactions(t *testing.T) { db, delete := openTestDB(t, nil) defer func() { testutil.Ok(t, db.Close()) delete() }() blockRange := DefaultOptions.BlockRanges[0] label := labels.FromStrings("foo", "bar") // Trigger a compaction to check that it was skipped and // no new blocks were created when compaction is disabled. db.DisableCompactions() app := db.Appender() for i := int64(0); i < 3; i++ { _, err := app.Add(label, i*blockRange, 0) testutil.Ok(t, err) _, err = app.Add(label, i*blockRange+1000, 0) testutil.Ok(t, err) } testutil.Ok(t, app.Commit()) select { case db.compactc <- struct{}{}: default: } m := &dto.Metric{} for x := 0; x < 10; x++ { db.metrics.compactionsSkipped.Write(m) if *m.Counter.Value > float64(0) { break } time.Sleep(10 * time.Millisecond) } testutil.Assert(t, *m.Counter.Value > float64(0), "No compaction was skipped after the set timeout.") testutil.Equals(t, 0, len(db.blocks)) // Enable the compaction, trigger it and check that the block is persisted. db.EnableCompactions() select { case db.compactc <- struct{}{}: default: } for x := 0; x < 100; x++ { if len(db.Blocks()) > 0 { break } time.Sleep(100 * time.Millisecond) } testutil.Assert(t, len(db.Blocks()) > 0, "No block was persisted after the set timeout.") } // TestCancelCompactions ensures that when the db is closed // any running compaction is cancelled to unblock closing the db. func TestCancelCompactions(t *testing.T) { tmpdir, err := ioutil.TempDir("", "testCancelCompaction") testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(tmpdir)) }() // Create some blocks to fall within the compaction range. createBlock(t, tmpdir, genSeries(10, 10000, 0, 1000)) createBlock(t, tmpdir, genSeries(10, 10000, 1000, 2000)) createBlock(t, tmpdir, genSeries(1, 1, 2000, 2001)) // The most recent block is ignored so can be e small one. // Copy the db so we have an exact copy to compare compaction times. tmpdirCopy := tmpdir + "Copy" err = fileutil.CopyDirs(tmpdir, tmpdirCopy) testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(tmpdirCopy)) }() // Measure the compaction time without interupting it. var timeCompactionUninterrupted time.Duration { db, err := Open(tmpdir, log.NewNopLogger(), nil, &Options{BlockRanges: []int64{1, 2000}}) testutil.Ok(t, err) testutil.Equals(t, 3, len(db.Blocks()), "initial block count mismatch") testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "initial compaction counter mismatch") db.compactc <- struct{}{} // Trigger a compaction. var start time.Time for prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.populatingBlocks) <= 0 { time.Sleep(3 * time.Millisecond) } start = time.Now() for prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran) != 1 { time.Sleep(3 * time.Millisecond) } timeCompactionUninterrupted = time.Since(start) testutil.Ok(t, db.Close()) } // Measure the compaction time when closing the db in the middle of compaction. { db, err := Open(tmpdirCopy, log.NewNopLogger(), nil, &Options{BlockRanges: []int64{1, 2000}}) testutil.Ok(t, err) testutil.Equals(t, 3, len(db.Blocks()), "initial block count mismatch") testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "initial compaction counter mismatch") db.compactc <- struct{}{} // Trigger a compaction. dbClosed := make(chan struct{}) for prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.populatingBlocks) <= 0 { time.Sleep(3 * time.Millisecond) } go func() { testutil.Ok(t, db.Close()) close(dbClosed) }() start := time.Now() <-dbClosed actT := time.Since(start) expT := time.Duration(timeCompactionUninterrupted / 2) // Closing the db in the middle of compaction should less than half the time. testutil.Assert(t, actT < expT, "closing the db took more than expected. exp: <%v, act: %v", expT, actT) } } // TestDeleteCompactionBlockAfterFailedReload ensures that a failed reload immediately after a compaction // deletes the resulting block to avoid creatings blocks with the same time range. func TestDeleteCompactionBlockAfterFailedReload(t *testing.T) { tests := map[string]func(*DB) int{ "Test Head Compaction": func(db *DB) int { rangeToTriggerCompaction := db.opts.BlockRanges[0]/2*3 - 1 defaultLabel := labels.FromStrings("foo", "bar") // Add some data to the head that is enough to trigger a compaction. app := db.Appender() _, err := app.Add(defaultLabel, 1, 0) testutil.Ok(t, err) _, err = app.Add(defaultLabel, 2, 0) testutil.Ok(t, err) _, err = app.Add(defaultLabel, 3+rangeToTriggerCompaction, 0) testutil.Ok(t, err) testutil.Ok(t, app.Commit()) return 0 }, "Test Block Compaction": func(db *DB) int { blocks := []*BlockMeta{ {MinTime: 0, MaxTime: 100}, {MinTime: 100, MaxTime: 150}, {MinTime: 150, MaxTime: 200}, } for _, m := range blocks { createBlock(t, db.Dir(), genSeries(1, 1, m.MinTime, m.MaxTime)) } testutil.Ok(t, db.reload()) testutil.Equals(t, len(blocks), len(db.Blocks()), "unexpected block count after a reload") return len(blocks) }, } for title, bootStrap := range tests { t.Run(title, func(t *testing.T) { db, delete := openTestDB(t, &Options{ BlockRanges: []int64{1, 100}, }) defer func() { testutil.Ok(t, db.Close()) delete() }() db.DisableCompactions() expBlocks := bootStrap(db) // Create a block that will trigger the reload to fail. blockPath := createBlock(t, db.Dir(), genSeries(1, 1, 200, 300)) lastBlockIndex := path.Join(blockPath, indexFilename) actBlocks, err := blockDirs(db.Dir()) testutil.Ok(t, err) testutil.Equals(t, expBlocks, len(actBlocks)-1) // -1 to exclude the corrupted block. testutil.Ok(t, os.RemoveAll(lastBlockIndex)) // Corrupt the block by removing the index file. testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.metrics.reloadsFailed), "initial 'failed db reload' count metrics mismatch") testutil.Equals(t, 0.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "initial `compactions` count metric mismatch") // Do the compaction and check the metrics. // Compaction should succeed, but the reload should fail and // the new block created from the compaction should be deleted. testutil.NotOk(t, db.compact()) testutil.Equals(t, 1.0, prom_testutil.ToFloat64(db.metrics.reloadsFailed), "'failed db reload' count metrics mismatch") testutil.Equals(t, 1.0, prom_testutil.ToFloat64(db.compactor.(*LeveledCompactor).metrics.ran), "`compaction` count metric mismatch") actBlocks, err = blockDirs(db.Dir()) testutil.Ok(t, err) testutil.Equals(t, expBlocks, len(actBlocks)-1, "block count should be the same as before the compaction") // -1 to exclude the corrupted block. }) } }