Merge branch 'beorn7/persistence' into beorn7/ingestion-tweaks

This commit is contained in:
beorn7 2015-03-18 19:09:31 +01:00
commit a075900f9a
7 changed files with 212 additions and 18 deletions

View File

@ -639,15 +639,12 @@ func evalScalarBinop(opType BinOpType,
case Mul:
return lhs * rhs
case Div:
if rhs != 0 {
return lhs / rhs
}
return clientmodel.SampleValue(math.Inf(int(rhs)))
return lhs / rhs
case Mod:
if rhs != 0 {
return clientmodel.SampleValue(int(lhs) % int(rhs))
}
return clientmodel.SampleValue(math.Inf(int(rhs)))
return clientmodel.SampleValue(math.NaN())
case EQ:
if lhs == rhs {
return 1
@ -693,15 +690,12 @@ func evalVectorBinop(opType BinOpType,
case Mul:
return lhs * rhs, true
case Div:
if rhs != 0 {
return lhs / rhs, true
}
return clientmodel.SampleValue(math.Inf(int(rhs))), true
return lhs / rhs, true
case Mod:
if rhs != 0 {
return clientmodel.SampleValue(int(lhs) % int(rhs)), true
}
return clientmodel.SampleValue(math.Inf(int(rhs))), true
return clientmodel.SampleValue(math.NaN()), true
case EQ:
if lhs == rhs {
return lhs, true

View File

@ -456,6 +456,50 @@ func ceilImpl(timestamp clientmodel.Timestamp, args []Node) interface{} {
return vector
}
// === exp(vector VectorNode) Vector ===
func expImpl(timestamp clientmodel.Timestamp, args []Node) interface{} {
n := args[0].(VectorNode)
vector := n.Eval(timestamp)
for _, el := range vector {
el.Metric.Delete(clientmodel.MetricNameLabel)
el.Value = clientmodel.SampleValue(math.Exp(float64(el.Value)))
}
return vector
}
// === ln(vector VectorNode) Vector ===
func lnImpl(timestamp clientmodel.Timestamp, args []Node) interface{} {
n := args[0].(VectorNode)
vector := n.Eval(timestamp)
for _, el := range vector {
el.Metric.Delete(clientmodel.MetricNameLabel)
el.Value = clientmodel.SampleValue(math.Log(float64(el.Value)))
}
return vector
}
// === log2(vector VectorNode) Vector ===
func log2Impl(timestamp clientmodel.Timestamp, args []Node) interface{} {
n := args[0].(VectorNode)
vector := n.Eval(timestamp)
for _, el := range vector {
el.Metric.Delete(clientmodel.MetricNameLabel)
el.Value = clientmodel.SampleValue(math.Log2(float64(el.Value)))
}
return vector
}
// === log10(vector VectorNode) Vector ===
func log10Impl(timestamp clientmodel.Timestamp, args []Node) interface{} {
n := args[0].(VectorNode)
vector := n.Eval(timestamp)
for _, el := range vector {
el.Metric.Delete(clientmodel.MetricNameLabel)
el.Value = clientmodel.SampleValue(math.Log10(float64(el.Value)))
}
return vector
}
// === deriv(node MatrixNode) Vector ===
func derivImpl(timestamp clientmodel.Timestamp, args []Node) interface{} {
matrixNode := args[0].(MatrixNode)
@ -598,6 +642,12 @@ var functions = map[string]*Function{
returnType: VectorType,
callFn: dropCommonLabelsImpl,
},
"exp": {
name: "exp",
argTypes: []ExprType{VectorType},
returnType: VectorType,
callFn: expImpl,
},
"floor": {
name: "floor",
argTypes: []ExprType{VectorType},
@ -610,6 +660,24 @@ var functions = map[string]*Function{
returnType: VectorType,
callFn: histogramQuantileImpl,
},
"ln": {
name: "ln",
argTypes: []ExprType{VectorType},
returnType: VectorType,
callFn: lnImpl,
},
"log10": {
name: "log10",
argTypes: []ExprType{VectorType},
returnType: VectorType,
callFn: log10Impl,
},
"log2": {
name: "log2",
argTypes: []ExprType{VectorType},
returnType: VectorType,
callFn: log2Impl,
},
"max_over_time": {
name: "max_over_time",
argTypes: []ExprType{MatrixType},

View File

@ -36,7 +36,7 @@ var (
testEvalTime = testStartTime.Add(testSampleInterval * 10)
fixturesPath = "fixtures"
reSample = regexp.MustCompile(`^(.*)(?: \=\>|:) (\-?\d+\.?\d*e?\d*|[+-]Inf|NaN) \@\[(\d+)\]$`)
reSample = regexp.MustCompile(`^(.*)(?: \=\>|:) (\-?\d+\.?\d*(?:e-?\d*)?|[+-]Inf|NaN) \@\[(\d+)\]$`)
minNormal = math.Float64frombits(0x0010000000000000) // The smallest positive normal value of type float64.
)
@ -1157,6 +1157,137 @@ func TestExpressions(t *testing.T) {
expr: `999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999`,
output: []string{`scalar: +Inf @[%v]`},
},
{
expr: `1 / 0`,
output: []string{`scalar: +Inf @[%v]`},
},
{
expr: `-1 / 0`,
output: []string{`scalar: -Inf @[%v]`},
},
{
expr: `0 / 0`,
output: []string{`scalar: NaN @[%v]`},
},
{
expr: `1 % 0`,
output: []string{`scalar: NaN @[%v]`},
},
{
expr: `http_requests{group="canary", instance="0", job="api-server"} / 0`,
output: []string{
`{group="canary", instance="0", job="api-server"} => +Inf @[%v]`,
},
},
{
expr: `-1 * http_requests{group="canary", instance="0", job="api-server"} / 0`,
output: []string{
`{group="canary", instance="0", job="api-server"} => -Inf @[%v]`,
},
},
{
expr: `0 * http_requests{group="canary", instance="0", job="api-server"} / 0`,
output: []string{
`{group="canary", instance="0", job="api-server"} => NaN @[%v]`,
},
},
{
expr: `0 * http_requests{group="canary", instance="0", job="api-server"} % 0`,
output: []string{
`{group="canary", instance="0", job="api-server"} => NaN @[%v]`,
},
},
{
expr: `exp(vector_matching_a)`,
output: []string{
`{l="x"} => 22026.465794806718 @[%v]`,
`{l="y"} => 485165195.4097903 @[%v]`,
},
},
{
expr: `exp(vector_matching_a - 10)`,
output: []string{
`{l="y"} => 22026.465794806718 @[%v]`,
`{l="x"} => 1 @[%v]`,
},
},
{
expr: `exp(vector_matching_a - 20)`,
output: []string{
`{l="x"} => 4.5399929762484854e-05 @[%v]`,
`{l="y"} => 1 @[%v]`,
},
},
{
expr: `ln(vector_matching_a)`,
output: []string{
`{l="x"} => 2.302585092994046 @[%v]`,
`{l="y"} => 2.995732273553991 @[%v]`,
},
},
{
expr: `ln(vector_matching_a - 10)`,
output: []string{
`{l="y"} => 2.302585092994046 @[%v]`,
`{l="x"} => -Inf @[%v]`,
},
},
{
expr: `ln(vector_matching_a - 20)`,
output: []string{
`{l="y"} => -Inf @[%v]`,
`{l="x"} => NaN @[%v]`,
},
},
{
expr: `exp(ln(vector_matching_a))`,
output: []string{
`{l="y"} => 20 @[%v]`,
`{l="x"} => 10 @[%v]`,
},
},
{
expr: `log2(vector_matching_a)`,
output: []string{
`{l="x"} => 3.3219280948873626 @[%v]`,
`{l="y"} => 4.321928094887363 @[%v]`,
},
},
{
expr: `log2(vector_matching_a - 10)`,
output: []string{
`{l="y"} => 3.3219280948873626 @[%v]`,
`{l="x"} => -Inf @[%v]`,
},
},
{
expr: `log2(vector_matching_a - 20)`,
output: []string{
`{l="x"} => NaN @[%v]`,
`{l="y"} => -Inf @[%v]`,
},
},
{
expr: `log10(vector_matching_a)`,
output: []string{
`{l="x"} => 1 @[%v]`,
`{l="y"} => 1.301029995663981 @[%v]`,
},
},
{
expr: `log10(vector_matching_a - 10)`,
output: []string{
`{l="y"} => 1 @[%v]`,
`{l="x"} => -Inf @[%v]`,
},
},
{
expr: `log10(vector_matching_a - 20)`,
output: []string{
`{l="x"} => NaN @[%v]`,
`{l="y"} => -Inf @[%v]`,
},
},
}
storage, closer := newTestStorage(t)

View File

@ -225,7 +225,9 @@ func (p *persistence) sanitizeSeries(
// Everything is consistent. We are good.
return fp, true
}
// If we are here, something's fishy.
// If we are here, we cannot be sure the series file is
// consistent with the checkpoint, so we have to take a closer
// look.
if s.headChunkClosed {
// This is the easy case as we don't have any chunks in
// heads.db. Treat this series as a freshly unarchived

View File

@ -911,7 +911,7 @@ func (p *persistence) dropAndPersistChunks(
// Otherwise, seek backwards to the beginning of its header and start
// copying everything from there into a new file. Then append the chunks
// to the new file.
_, err = f.Seek(-(chunkHeaderLen), os.SEEK_CUR)
_, err = f.Seek(-chunkHeaderLen, os.SEEK_CUR)
if err != nil {
return
}

View File

@ -32,7 +32,6 @@ const (
chunkDescEvictionFactor = 10
headChunkTimeout = time.Hour // Close head chunk if not touched for that long.
)
// fingerprintSeriesPair pairs a fingerprint with a memorySeries pointer.
@ -168,7 +167,7 @@ type memorySeries struct {
// appended.
headChunkUsedByIterator bool
// Whether the series is inconsistent with the last checkpoint in a way
// that would require a desk seek during crash recovery.
// that would require a disk seek during crash recovery.
dirty bool
}
@ -233,7 +232,7 @@ func (s *memorySeries) add(v *metric.SamplePair) int {
}
// maybeCloseHeadChunk closes the head chunk if it has not been touched for the
// duration of headChunkTimeout. It returns wether the head chunk was closed.
// duration of headChunkTimeout. It returns whether the head chunk was closed.
// If the head chunk is already closed, the method is a no-op and returns false.
//
// The caller must have locked the fingerprint of the series.

View File

@ -761,8 +761,8 @@ func (s *memorySeriesStorage) maintainMemorySeries(
}
// writeMemorySeries (re-)writes a memory series file. While doing so, it drops
// chunks older than beforeTime from both, the series file (if it exists) as
// well as from memory. The provided chunksToPersist are appended to the newly
// chunks older than beforeTime from both the series file (if it exists) as well
// as from memory. The provided chunksToPersist are appended to the newly
// written series file. If no chunks need to be purged, but chunksToPersist is
// not empty, those chunks are simply appended to the series file. If the series
// contains no chunks after dropping old chunks, it is purged entirely. In that