Merge pull request #14629 from krajorama/fix-to-encoded-chunks
Fix ToEncodedChunks minT for recoded chunks
This commit is contained in:
commit
cf62fb5c44
|
@ -1094,7 +1094,7 @@ func (s *memSeries) insert(t int64, v float64, chunkDiskMapper *chunks.ChunkDisk
|
||||||
chunkCreated = true
|
chunkCreated = true
|
||||||
}
|
}
|
||||||
|
|
||||||
ok := c.chunk.Insert(t, v)
|
ok := c.chunk.Insert(t, v, nil, nil)
|
||||||
if ok {
|
if ok {
|
||||||
if chunkCreated || t < c.minTime {
|
if chunkCreated || t < c.minTime {
|
||||||
c.minTime = t
|
c.minTime = t
|
||||||
|
|
|
@ -17,6 +17,7 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"sort"
|
"sort"
|
||||||
|
|
||||||
|
"github.com/prometheus/prometheus/model/histogram"
|
||||||
"github.com/prometheus/prometheus/tsdb/chunkenc"
|
"github.com/prometheus/prometheus/tsdb/chunkenc"
|
||||||
|
|
||||||
"github.com/oklog/ulid"
|
"github.com/oklog/ulid"
|
||||||
|
@ -39,13 +40,13 @@ func NewOOOChunk() *OOOChunk {
|
||||||
|
|
||||||
// Insert inserts the sample such that order is maintained.
|
// Insert inserts the sample such that order is maintained.
|
||||||
// Returns false if insert was not possible due to the same timestamp already existing.
|
// Returns false if insert was not possible due to the same timestamp already existing.
|
||||||
func (o *OOOChunk) Insert(t int64, v float64) bool {
|
func (o *OOOChunk) Insert(t int64, v float64, h *histogram.Histogram, fh *histogram.FloatHistogram) bool {
|
||||||
// Although out-of-order samples can be out-of-order amongst themselves, we
|
// Although out-of-order samples can be out-of-order amongst themselves, we
|
||||||
// are opinionated and expect them to be usually in-order meaning we could
|
// are opinionated and expect them to be usually in-order meaning we could
|
||||||
// try to append at the end first if the new timestamp is higher than the
|
// try to append at the end first if the new timestamp is higher than the
|
||||||
// last known timestamp.
|
// last known timestamp.
|
||||||
if len(o.samples) == 0 || t > o.samples[len(o.samples)-1].t {
|
if len(o.samples) == 0 || t > o.samples[len(o.samples)-1].t {
|
||||||
o.samples = append(o.samples, sample{t, v, nil, nil})
|
o.samples = append(o.samples, sample{t, v, h, fh})
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -54,7 +55,7 @@ func (o *OOOChunk) Insert(t int64, v float64) bool {
|
||||||
|
|
||||||
if i >= len(o.samples) {
|
if i >= len(o.samples) {
|
||||||
// none found. append it at the end
|
// none found. append it at the end
|
||||||
o.samples = append(o.samples, sample{t, v, nil, nil})
|
o.samples = append(o.samples, sample{t, v, h, fh})
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -66,7 +67,7 @@ func (o *OOOChunk) Insert(t int64, v float64) bool {
|
||||||
// Expand length by 1 to make room. use a zero sample, we will overwrite it anyway.
|
// Expand length by 1 to make room. use a zero sample, we will overwrite it anyway.
|
||||||
o.samples = append(o.samples, sample{})
|
o.samples = append(o.samples, sample{})
|
||||||
copy(o.samples[i+1:], o.samples[i:])
|
copy(o.samples[i+1:], o.samples[i:])
|
||||||
o.samples[i] = sample{t, v, nil, nil}
|
o.samples[i] = sample{t, v, h, fh}
|
||||||
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
@ -142,9 +143,9 @@ func (o *OOOChunk) ToEncodedChunks(mint, maxt int64) (chks []memChunk, err error
|
||||||
if newChunk != nil { // A new chunk was allocated.
|
if newChunk != nil { // A new chunk was allocated.
|
||||||
if !recoded {
|
if !recoded {
|
||||||
chks = append(chks, memChunk{chunk, cmint, cmaxt, nil})
|
chks = append(chks, memChunk{chunk, cmint, cmaxt, nil})
|
||||||
|
cmint = s.t
|
||||||
}
|
}
|
||||||
chunk = newChunk
|
chunk = newChunk
|
||||||
cmint = s.t
|
|
||||||
}
|
}
|
||||||
case chunkenc.EncFloatHistogram:
|
case chunkenc.EncFloatHistogram:
|
||||||
// Ignoring ok is ok, since we don't want to compare to the wrong previous appender anyway.
|
// Ignoring ok is ok, since we don't want to compare to the wrong previous appender anyway.
|
||||||
|
@ -157,9 +158,9 @@ func (o *OOOChunk) ToEncodedChunks(mint, maxt int64) (chks []memChunk, err error
|
||||||
if newChunk != nil { // A new chunk was allocated.
|
if newChunk != nil { // A new chunk was allocated.
|
||||||
if !recoded {
|
if !recoded {
|
||||||
chks = append(chks, memChunk{chunk, cmint, cmaxt, nil})
|
chks = append(chks, memChunk{chunk, cmint, cmaxt, nil})
|
||||||
|
cmint = s.t
|
||||||
}
|
}
|
||||||
chunk = newChunk
|
chunk = newChunk
|
||||||
cmint = s.t
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
cmaxt = s.t
|
cmaxt = s.t
|
||||||
|
|
|
@ -14,8 +14,14 @@
|
||||||
package tsdb
|
package tsdb
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"math"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"github.com/prometheus/prometheus/model/histogram"
|
||||||
|
"github.com/prometheus/prometheus/storage"
|
||||||
|
"github.com/prometheus/prometheus/tsdb/chunkenc"
|
||||||
|
"github.com/prometheus/prometheus/tsdb/tsdbutil"
|
||||||
|
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -52,7 +58,7 @@ func TestOOOInsert(t *testing.T) {
|
||||||
chunk := NewOOOChunk()
|
chunk := NewOOOChunk()
|
||||||
chunk.samples = makeEvenSampleSlice(numPreExisting)
|
chunk.samples = makeEvenSampleSlice(numPreExisting)
|
||||||
newSample := samplify(valOdd(insertPos))
|
newSample := samplify(valOdd(insertPos))
|
||||||
chunk.Insert(newSample.t, newSample.f)
|
chunk.Insert(newSample.t, newSample.f, nil, nil)
|
||||||
|
|
||||||
var expSamples []sample
|
var expSamples []sample
|
||||||
// Our expected new samples slice, will be first the original samples.
|
// Our expected new samples slice, will be first the original samples.
|
||||||
|
@ -83,7 +89,7 @@ func TestOOOInsertDuplicate(t *testing.T) {
|
||||||
dupSample := chunk.samples[dupPos]
|
dupSample := chunk.samples[dupPos]
|
||||||
dupSample.f = 0.123
|
dupSample.f = 0.123
|
||||||
|
|
||||||
ok := chunk.Insert(dupSample.t, dupSample.f)
|
ok := chunk.Insert(dupSample.t, dupSample.f, nil, nil)
|
||||||
|
|
||||||
expSamples := makeEvenSampleSlice(num) // We expect no change.
|
expSamples := makeEvenSampleSlice(num) // We expect no change.
|
||||||
require.False(t, ok)
|
require.False(t, ok)
|
||||||
|
@ -91,3 +97,136 @@ func TestOOOInsertDuplicate(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type chunkVerify struct {
|
||||||
|
encoding chunkenc.Encoding
|
||||||
|
minTime int64
|
||||||
|
maxTime int64
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestOOOChunks_ToEncodedChunks(t *testing.T) {
|
||||||
|
h1 := tsdbutil.GenerateTestHistogram(1)
|
||||||
|
// Make h2 appendable but with more buckets, to trigger recoding.
|
||||||
|
h2 := h1.Copy()
|
||||||
|
h2.PositiveSpans = append(h2.PositiveSpans, histogram.Span{Offset: 1, Length: 1})
|
||||||
|
h2.PositiveBuckets = append(h2.PositiveBuckets, 12)
|
||||||
|
|
||||||
|
testCases := map[string]struct {
|
||||||
|
samples []sample
|
||||||
|
expectedCounterResets []histogram.CounterResetHint
|
||||||
|
expectedChunks []chunkVerify
|
||||||
|
}{
|
||||||
|
"empty": {
|
||||||
|
samples: []sample{},
|
||||||
|
},
|
||||||
|
"has floats": {
|
||||||
|
samples: []sample{
|
||||||
|
{t: 1000, f: 43.0},
|
||||||
|
{t: 1100, f: 42.0},
|
||||||
|
},
|
||||||
|
expectedCounterResets: []histogram.CounterResetHint{histogram.UnknownCounterReset, histogram.UnknownCounterReset},
|
||||||
|
expectedChunks: []chunkVerify{
|
||||||
|
{encoding: chunkenc.EncXOR, minTime: 1000, maxTime: 1100},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"mix of floats and histograms": {
|
||||||
|
samples: []sample{
|
||||||
|
{t: 1000, f: 43.0},
|
||||||
|
{t: 1100, h: h1},
|
||||||
|
{t: 1200, f: 42.0},
|
||||||
|
},
|
||||||
|
expectedCounterResets: []histogram.CounterResetHint{histogram.UnknownCounterReset, histogram.UnknownCounterReset, histogram.UnknownCounterReset},
|
||||||
|
expectedChunks: []chunkVerify{
|
||||||
|
{encoding: chunkenc.EncXOR, minTime: 1000, maxTime: 1000},
|
||||||
|
{encoding: chunkenc.EncHistogram, minTime: 1100, maxTime: 1100},
|
||||||
|
{encoding: chunkenc.EncXOR, minTime: 1200, maxTime: 1200},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"has a counter reset": {
|
||||||
|
samples: []sample{
|
||||||
|
{t: 1000, h: h2},
|
||||||
|
{t: 1100, h: h1},
|
||||||
|
},
|
||||||
|
expectedCounterResets: []histogram.CounterResetHint{histogram.UnknownCounterReset, histogram.CounterReset},
|
||||||
|
expectedChunks: []chunkVerify{
|
||||||
|
{encoding: chunkenc.EncHistogram, minTime: 1000, maxTime: 1000},
|
||||||
|
{encoding: chunkenc.EncHistogram, minTime: 1100, maxTime: 1100},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
"has a recoded histogram": { // Regression test for wrong minT, maxT in histogram recoding.
|
||||||
|
samples: []sample{
|
||||||
|
{t: 0, h: h1},
|
||||||
|
{t: 1, h: h2},
|
||||||
|
},
|
||||||
|
expectedCounterResets: []histogram.CounterResetHint{histogram.UnknownCounterReset, histogram.NotCounterReset},
|
||||||
|
expectedChunks: []chunkVerify{
|
||||||
|
{encoding: chunkenc.EncHistogram, minTime: 0, maxTime: 1},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, tc := range testCases {
|
||||||
|
t.Run(name, func(t *testing.T) {
|
||||||
|
// Sanity check.
|
||||||
|
require.Equal(t, len(tc.samples), len(tc.expectedCounterResets), "number of samples and counter resets")
|
||||||
|
|
||||||
|
oooChunk := OOOChunk{}
|
||||||
|
for _, s := range tc.samples {
|
||||||
|
switch s.Type() {
|
||||||
|
case chunkenc.ValFloat:
|
||||||
|
oooChunk.Insert(s.t, s.f, nil, nil)
|
||||||
|
case chunkenc.ValHistogram:
|
||||||
|
oooChunk.Insert(s.t, 0, s.h.Copy(), nil)
|
||||||
|
case chunkenc.ValFloatHistogram:
|
||||||
|
oooChunk.Insert(s.t, 0, nil, s.fh.Copy())
|
||||||
|
default:
|
||||||
|
t.Fatalf("unexpected sample type %d", s.Type())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
chunks, err := oooChunk.ToEncodedChunks(math.MinInt64, math.MaxInt64)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, len(tc.expectedChunks), len(chunks), "number of chunks")
|
||||||
|
sampleIndex := 0
|
||||||
|
for i, c := range chunks {
|
||||||
|
require.Equal(t, tc.expectedChunks[i].encoding, c.chunk.Encoding(), "chunk %d encoding", i)
|
||||||
|
require.Equal(t, tc.expectedChunks[i].minTime, c.minTime, "chunk %d minTime", i)
|
||||||
|
require.Equal(t, tc.expectedChunks[i].maxTime, c.maxTime, "chunk %d maxTime", i)
|
||||||
|
samples, err := storage.ExpandSamples(c.chunk.Iterator(nil), newSample)
|
||||||
|
require.GreaterOrEqual(t, len(tc.samples)-sampleIndex, len(samples), "too many samples in chunk %d expected less than %d", i, len(tc.samples)-sampleIndex)
|
||||||
|
require.NoError(t, err)
|
||||||
|
if len(samples) == 0 {
|
||||||
|
// Ignore empty chunks.
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
switch c.chunk.Encoding() {
|
||||||
|
case chunkenc.EncXOR:
|
||||||
|
for j, s := range samples {
|
||||||
|
require.Equal(t, chunkenc.ValFloat, s.Type())
|
||||||
|
// XOR chunks don't have counter reset hints, so we shouldn't expect anything else than UnknownCounterReset.
|
||||||
|
require.Equal(t, histogram.UnknownCounterReset, tc.expectedCounterResets[sampleIndex+j], "sample reset hint %d", sampleIndex+j)
|
||||||
|
require.Equal(t, tc.samples[sampleIndex+j].f, s.F(), "sample %d", sampleIndex+j)
|
||||||
|
}
|
||||||
|
case chunkenc.EncHistogram:
|
||||||
|
for j, s := range samples {
|
||||||
|
require.Equal(t, chunkenc.ValHistogram, s.Type())
|
||||||
|
require.Equal(t, tc.expectedCounterResets[sampleIndex+j], s.H().CounterResetHint, "sample reset hint %d", sampleIndex+j)
|
||||||
|
compareTo := tc.samples[sampleIndex+j].h.Copy()
|
||||||
|
compareTo.CounterResetHint = tc.expectedCounterResets[sampleIndex+j]
|
||||||
|
require.Equal(t, compareTo, s.H().Compact(0), "sample %d", sampleIndex+j)
|
||||||
|
}
|
||||||
|
case chunkenc.EncFloatHistogram:
|
||||||
|
for j, s := range samples {
|
||||||
|
require.Equal(t, chunkenc.ValFloatHistogram, s.Type())
|
||||||
|
require.Equal(t, tc.expectedCounterResets[sampleIndex+j], s.FH().CounterResetHint, "sample reset hint %d", sampleIndex+j)
|
||||||
|
compareTo := tc.samples[sampleIndex+j].fh.Copy()
|
||||||
|
compareTo.CounterResetHint = tc.expectedCounterResets[sampleIndex+j]
|
||||||
|
require.Equal(t, compareTo, s.FH().Compact(0), "sample %d", sampleIndex+j)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sampleIndex += len(samples)
|
||||||
|
}
|
||||||
|
require.Equal(t, len(tc.samples), sampleIndex, "number of samples")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in New Issue