// Copyright 2017 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package index import ( "io/ioutil" "math/rand" "os" "path/filepath" "sort" "testing" "github.com/pkg/errors" "github.com/prometheus/tsdb/chunkenc" "github.com/prometheus/tsdb/chunks" "github.com/prometheus/tsdb/encoding" "github.com/prometheus/tsdb/labels" "github.com/prometheus/tsdb/testutil" ) type series struct { l labels.Labels chunks []chunks.Meta } type mockIndex struct { series map[uint64]series labelIndex map[string][]string postings map[labels.Label][]uint64 symbols map[string]struct{} } func newMockIndex() mockIndex { ix := mockIndex{ series: make(map[uint64]series), labelIndex: make(map[string][]string), postings: make(map[labels.Label][]uint64), symbols: make(map[string]struct{}), } return ix } func (m mockIndex) Symbols() (map[string]struct{}, error) { return m.symbols, nil } func (m mockIndex) AddSeries(ref uint64, l labels.Labels, chunks ...chunks.Meta) error { if _, ok := m.series[ref]; ok { return errors.Errorf("series with reference %d already added", ref) } for _, lbl := range l { m.symbols[lbl.Name] = struct{}{} m.symbols[lbl.Value] = struct{}{} } s := series{l: l} // Actual chunk data is not stored in the index. for _, c := range chunks { c.Chunk = nil s.chunks = append(s.chunks, c) } m.series[ref] = s return nil } func (m mockIndex) WriteLabelIndex(names []string, values []string) error { // TODO support composite indexes if len(names) != 1 { return errors.New("composite indexes not supported yet") } sort.Strings(values) m.labelIndex[names[0]] = values return nil } func (m mockIndex) WritePostings(name, value string, it Postings) error { l := labels.Label{Name: name, Value: value} if _, ok := m.postings[l]; ok { return errors.Errorf("postings for %s already added", l) } ep, err := ExpandPostings(it) if err != nil { return err } m.postings[l] = ep return nil } func (m mockIndex) Close() error { return nil } func (m mockIndex) LabelValues(names ...string) (StringTuples, error) { // TODO support composite indexes if len(names) != 1 { return nil, errors.New("composite indexes not supported yet") } return NewStringTuples(m.labelIndex[names[0]], 1) } func (m mockIndex) Postings(name, value string) (Postings, error) { l := labels.Label{Name: name, Value: value} return NewListPostings(m.postings[l]), nil } func (m mockIndex) SortedPostings(p Postings) Postings { ep, err := ExpandPostings(p) if err != nil { return ErrPostings(errors.Wrap(err, "expand postings")) } sort.Slice(ep, func(i, j int) bool { return labels.Compare(m.series[ep[i]].l, m.series[ep[j]].l) < 0 }) return NewListPostings(ep) } func (m mockIndex) Series(ref uint64, lset *labels.Labels, chks *[]chunks.Meta) error { s, ok := m.series[ref] if !ok { return errors.New("not found") } *lset = append((*lset)[:0], s.l...) *chks = append((*chks)[:0], s.chunks...) return nil } func (m mockIndex) LabelIndices() ([][]string, error) { res := make([][]string, 0, len(m.labelIndex)) for k := range m.labelIndex { res = append(res, []string{k}) } return res, nil } func TestIndexRW_Create_Open(t *testing.T) { dir, err := ioutil.TempDir("", "test_index_create") testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(dir)) }() fn := filepath.Join(dir, indexFilename) // An empty index must still result in a readable file. iw, err := NewWriter(fn) testutil.Ok(t, err) testutil.Ok(t, iw.Close()) ir, err := NewFileReader(fn) testutil.Ok(t, err) testutil.Ok(t, ir.Close()) // Modify magic header must cause open to fail. f, err := os.OpenFile(fn, os.O_WRONLY, 0666) testutil.Ok(t, err) _, err = f.WriteAt([]byte{0, 0}, 0) testutil.Ok(t, err) f.Close() _, err = NewFileReader(dir) testutil.NotOk(t, err) } func TestIndexRW_Postings(t *testing.T) { dir, err := ioutil.TempDir("", "test_index_postings") testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(dir)) }() fn := filepath.Join(dir, indexFilename) iw, err := NewWriter(fn) testutil.Ok(t, err) series := []labels.Labels{ labels.FromStrings("a", "1", "b", "1"), labels.FromStrings("a", "1", "b", "2"), labels.FromStrings("a", "1", "b", "3"), labels.FromStrings("a", "1", "b", "4"), } err = iw.AddSymbols(map[string]struct{}{ "a": {}, "b": {}, "1": {}, "2": {}, "3": {}, "4": {}, }) testutil.Ok(t, err) // Postings lists are only written if a series with the respective // reference was added before. testutil.Ok(t, iw.AddSeries(1, series[0])) testutil.Ok(t, iw.AddSeries(2, series[1])) testutil.Ok(t, iw.AddSeries(3, series[2])) testutil.Ok(t, iw.AddSeries(4, series[3])) err = iw.WritePostings("a", "1", newListPostings(1, 2, 3, 4)) testutil.Ok(t, err) testutil.Ok(t, iw.Close()) ir, err := NewFileReader(fn) testutil.Ok(t, err) p, err := ir.Postings("a", "1") testutil.Ok(t, err) var l labels.Labels var c []chunks.Meta for i := 0; p.Next(); i++ { err := ir.Series(p.At(), &l, &c) testutil.Ok(t, err) testutil.Equals(t, 0, len(c)) testutil.Equals(t, series[i], l) } testutil.Ok(t, p.Err()) testutil.Ok(t, ir.Close()) } func TestPersistence_index_e2e(t *testing.T) { dir, err := ioutil.TempDir("", "test_persistence_e2e") testutil.Ok(t, err) defer func() { testutil.Ok(t, os.RemoveAll(dir)) }() lbls, err := labels.ReadLabels(filepath.Join("..", "testdata", "20kseries.json"), 20000) testutil.Ok(t, err) // Sort labels as the index writer expects series in sorted order. sort.Sort(labels.Slice(lbls)) symbols := map[string]struct{}{} for _, lset := range lbls { for _, l := range lset { symbols[l.Name] = struct{}{} symbols[l.Value] = struct{}{} } } var input indexWriterSeriesSlice // Generate ChunkMetas for every label set. for i, lset := range lbls { var metas []chunks.Meta for j := 0; j <= (i % 20); j++ { metas = append(metas, chunks.Meta{ MinTime: int64(j * 10000), MaxTime: int64((j + 1) * 10000), Ref: rand.Uint64(), Chunk: chunkenc.NewXORChunk(), }) } input = append(input, &indexWriterSeries{ labels: lset, chunks: metas, }) } iw, err := NewWriter(filepath.Join(dir, indexFilename)) testutil.Ok(t, err) testutil.Ok(t, iw.AddSymbols(symbols)) // Population procedure as done by compaction. var ( postings = NewMemPostings() values = map[string]map[string]struct{}{} ) mi := newMockIndex() for i, s := range input { err = iw.AddSeries(uint64(i), s.labels, s.chunks...) testutil.Ok(t, err) testutil.Ok(t, mi.AddSeries(uint64(i), s.labels, s.chunks...)) for _, l := range s.labels { valset, ok := values[l.Name] if !ok { valset = map[string]struct{}{} values[l.Name] = valset } valset[l.Value] = struct{}{} } postings.Add(uint64(i), s.labels) } for k, v := range values { var vals []string for e := range v { vals = append(vals, e) } sort.Strings(vals) testutil.Ok(t, iw.WriteLabelIndex([]string{k}, vals)) testutil.Ok(t, mi.WriteLabelIndex([]string{k}, vals)) } all := make([]uint64, len(lbls)) for i := range all { all[i] = uint64(i) } err = iw.WritePostings("", "", newListPostings(all...)) testutil.Ok(t, err) testutil.Ok(t, mi.WritePostings("", "", newListPostings(all...))) for n, e := range postings.m { for v := range e { err = iw.WritePostings(n, v, postings.Get(n, v)) testutil.Ok(t, err) mi.WritePostings(n, v, postings.Get(n, v)) } } err = iw.Close() testutil.Ok(t, err) ir, err := NewFileReader(filepath.Join(dir, indexFilename)) testutil.Ok(t, err) for p := range mi.postings { gotp, err := ir.Postings(p.Name, p.Value) testutil.Ok(t, err) expp, err := mi.Postings(p.Name, p.Value) testutil.Ok(t, err) var lset, explset labels.Labels var chks, expchks []chunks.Meta for gotp.Next() { testutil.Assert(t, expp.Next() == true, "") ref := gotp.At() err := ir.Series(ref, &lset, &chks) testutil.Ok(t, err) err = mi.Series(expp.At(), &explset, &expchks) testutil.Ok(t, err) testutil.Equals(t, explset, lset) testutil.Equals(t, expchks, chks) } testutil.Assert(t, expp.Next() == false, "") testutil.Ok(t, gotp.Err()) } for k, v := range mi.labelIndex { tplsExp, err := NewStringTuples(v, 1) testutil.Ok(t, err) tplsRes, err := ir.LabelValues(k) testutil.Ok(t, err) testutil.Equals(t, tplsExp.Len(), tplsRes.Len()) for i := 0; i < tplsExp.Len(); i++ { strsExp, err := tplsExp.At(i) testutil.Ok(t, err) strsRes, err := tplsRes.At(i) testutil.Ok(t, err) testutil.Equals(t, strsExp, strsRes) } } gotSymbols, err := ir.Symbols() testutil.Ok(t, err) testutil.Equals(t, len(mi.symbols), len(gotSymbols)) for s := range mi.symbols { _, ok := gotSymbols[s] testutil.Assert(t, ok, "") } testutil.Ok(t, ir.Close()) } func TestDecbufUvariantWithInvalidBuffer(t *testing.T) { b := realByteSlice([]byte{0x81, 0x81, 0x81, 0x81, 0x81, 0x81}) db := encoding.NewDecbufUvarintAt(b, 0, castagnoliTable) testutil.NotOk(t, db.Err()) } func TestReaderWithInvalidBuffer(t *testing.T) { b := realByteSlice([]byte{0x81, 0x81, 0x81, 0x81, 0x81, 0x81}) _, err := NewReader(b) testutil.NotOk(t, err) } // TestNewFileReaderErrorNoOpenFiles ensures that in case of an error no file remains open. func TestNewFileReaderErrorNoOpenFiles(t *testing.T) { dir := testutil.NewTemporaryDirectory("block", t) idxName := filepath.Join(dir.Path(), "index") err := ioutil.WriteFile(idxName, []byte("corrupted contents"), 0644) testutil.Ok(t, err) _, err = NewFileReader(idxName) testutil.NotOk(t, err) // dir.Close will fail on Win if idxName fd is not closed on error path. dir.Close() }