diff --git a/cmd/promtool/tsdb.go b/cmd/promtool/tsdb.go index 369d26a4f..dfcfc320c 100644 --- a/cmd/promtool/tsdb.go +++ b/cmd/promtool/tsdb.go @@ -41,8 +41,6 @@ import ( tsdb_errors "github.com/prometheus/prometheus/tsdb/errors" ) -var merr tsdb_errors.MultiError - const timeDelta = 30000 type writeBenchmark struct { @@ -349,9 +347,7 @@ func listBlocks(path string, humanReadable bool) error { return err } defer func() { - merr.Add(err) - merr.Add(db.Close()) - err = merr.Err() + err = tsdb_errors.NewMulti(err, db.Close()).Err() }() blocks, err := db.Blocks() if err != nil { @@ -429,9 +425,7 @@ func analyzeBlock(path, blockID string, limit int) error { return err } defer func() { - merr.Add(err) - merr.Add(db.Close()) - err = merr.Err() + err = tsdb_errors.NewMulti(err, db.Close()).Err() }() meta := block.Meta() @@ -579,9 +573,7 @@ func dumpSamples(path string, mint, maxt int64) (err error) { return err } defer func() { - merr.Add(err) - merr.Add(db.Close()) - err = merr.Err() + err = tsdb_errors.NewMulti(err, db.Close()).Err() }() q, err := db.Querier(context.TODO(), mint, maxt) if err != nil { @@ -605,11 +597,7 @@ func dumpSamples(path string, mint, maxt int64) (err error) { } if ws := ss.Warnings(); len(ws) > 0 { - var merr tsdb_errors.MultiError - for _, w := range ws { - merr.Add(w) - } - return merr.Err() + return tsdb_errors.NewMulti(ws...).Err() } if ss.Err() != nil { diff --git a/storage/fanout.go b/storage/fanout.go index 82e65e91b..4bc3db12d 100644 --- a/storage/fanout.go +++ b/storage/fanout.go @@ -80,8 +80,7 @@ func (f *fanout) Querier(ctx context.Context, mint, maxt int64) (Querier, error) querier, err := storage.Querier(ctx, mint, maxt) if err != nil { // Close already open Queriers, append potential errors to returned error. - errs := tsdb_errors.MultiError{err} - errs.Add(primary.Close()) + errs := tsdb_errors.NewMulti(err, primary.Close()) for _, q := range secondaries { errs.Add(q.Close()) } @@ -103,8 +102,7 @@ func (f *fanout) ChunkQuerier(ctx context.Context, mint, maxt int64) (ChunkQueri querier, err := storage.ChunkQuerier(ctx, mint, maxt) if err != nil { // Close already open Queriers, append potential errors to returned error. - errs := tsdb_errors.MultiError{err} - errs.Add(primary.Close()) + errs := tsdb_errors.NewMulti(err, primary.Close()) for _, q := range secondaries { errs.Add(q.Close()) } @@ -130,8 +128,7 @@ func (f *fanout) Appender(ctx context.Context) Appender { // Close closes the storage and all its underlying resources. func (f *fanout) Close() error { - errs := tsdb_errors.MultiError{} - errs.Add(f.primary.Close()) + errs := tsdb_errors.NewMulti(f.primary.Close()) for _, s := range f.secondaries { errs.Add(s.Close()) } diff --git a/storage/merge.go b/storage/merge.go index 1670a119a..27e701883 100644 --- a/storage/merge.go +++ b/storage/merge.go @@ -248,7 +248,7 @@ func (q *mergeGenericQuerier) LabelNames() ([]string, Warnings, error) { // Close releases the resources of the generic querier. func (q *mergeGenericQuerier) Close() error { - errs := tsdb_errors.MultiError{} + errs := tsdb_errors.NewMulti() for _, querier := range q.queriers { if err := querier.Close(); err != nil { errs.Add(err) @@ -534,11 +534,9 @@ func (c *chainSampleIterator) Next() bool { } func (c *chainSampleIterator) Err() error { - var errs tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for _, iter := range c.iterators { - if err := iter.Err(); err != nil { - errs.Add(err) - } + errs.Add(iter.Err()) } return errs.Err() } @@ -681,11 +679,9 @@ func (c *compactChunkIterator) Next() bool { } func (c *compactChunkIterator) Err() error { - var errs tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for _, iter := range c.iterators { - if err := iter.Err(); err != nil { - errs.Add(err) - } + errs.Add(iter.Err()) } errs.Add(c.err) return errs.Err() diff --git a/tsdb/block.go b/tsdb/block.go index e177c112a..3ec226197 100644 --- a/tsdb/block.go +++ b/tsdb/block.go @@ -226,19 +226,14 @@ func writeMetaFile(logger log.Logger, dir string, meta *BlockMeta) (int64, error return 0, err } - var merr tsdb_errors.MultiError n, err := f.Write(jsonMeta) if err != nil { - merr.Add(err) - merr.Add(f.Close()) - return 0, merr.Err() + return 0, tsdb_errors.NewMulti(err, f.Close()).Err() } // Force the kernel to persist the file on disk to avoid data loss if the host crashes. if err := f.Sync(); err != nil { - merr.Add(err) - merr.Add(f.Close()) - return 0, merr.Err() + return 0, tsdb_errors.NewMulti(err, f.Close()).Err() } if err := f.Close(); err != nil { return 0, err @@ -280,10 +275,7 @@ func OpenBlock(logger log.Logger, dir string, pool chunkenc.Pool) (pb *Block, er var closers []io.Closer defer func() { if err != nil { - var merr tsdb_errors.MultiError - merr.Add(err) - merr.Add(closeAll(closers)) - err = merr.Err() + err = tsdb_errors.NewMulti(err, tsdb_errors.CloseAll(closers)).Err() } }() meta, sizeMeta, err := readMetaFile(dir) @@ -333,13 +325,11 @@ func (pb *Block) Close() error { pb.pendingReaders.Wait() - var merr tsdb_errors.MultiError - - merr.Add(pb.chunkr.Close()) - merr.Add(pb.indexr.Close()) - merr.Add(pb.tombstones.Close()) - - return merr.Err() + return tsdb_errors.NewMulti( + pb.chunkr.Close(), + pb.indexr.Close(), + pb.tombstones.Close(), + ).Err() } func (pb *Block) String() string { diff --git a/tsdb/chunks/chunks.go b/tsdb/chunks/chunks.go index af781e259..11417c38c 100644 --- a/tsdb/chunks/chunks.go +++ b/tsdb/chunks/chunks.go @@ -230,14 +230,13 @@ func cutSegmentFile(dirFile *os.File, magicNumber uint32, chunksFormat byte, all } defer func() { if returnErr != nil { - var merr tsdb_errors.MultiError - merr.Add(returnErr) + errs := tsdb_errors.NewMulti(returnErr) if f != nil { - merr.Add(f.Close()) + errs.Add(f.Close()) } // Calling RemoveAll on a non-existent file does not return error. - merr.Add(os.RemoveAll(ptmp)) - returnErr = merr.Err() + errs.Add(os.RemoveAll(ptmp)) + returnErr = errs.Err() } }() if allocSize > 0 { @@ -463,16 +462,16 @@ func NewDirReader(dir string, pool chunkenc.Pool) (*Reader, error) { } var ( - bs []ByteSlice - cs []io.Closer - merr tsdb_errors.MultiError + bs []ByteSlice + cs []io.Closer ) for _, fn := range files { f, err := fileutil.OpenMmapFile(fn) if err != nil { - merr.Add(errors.Wrap(err, "mmap files")) - merr.Add(closeAll(cs)) - return nil, merr + return nil, tsdb_errors.NewMulti( + errors.Wrap(err, "mmap files"), + tsdb_errors.CloseAll(cs), + ).Err() } cs = append(cs, f) bs = append(bs, realByteSlice(f.Bytes())) @@ -480,15 +479,16 @@ func NewDirReader(dir string, pool chunkenc.Pool) (*Reader, error) { reader, err := newReader(bs, cs, pool) if err != nil { - merr.Add(err) - merr.Add(closeAll(cs)) - return nil, merr + return nil, tsdb_errors.NewMulti( + err, + tsdb_errors.CloseAll(cs), + ).Err() } return reader, nil } func (s *Reader) Close() error { - return closeAll(s.cs) + return tsdb_errors.CloseAll(s.cs) } // Size returns the size of the chunks. @@ -588,12 +588,3 @@ func sequenceFiles(dir string) ([]string, error) { } return res, nil } - -func closeAll(cs []io.Closer) error { - var merr tsdb_errors.MultiError - - for _, c := range cs { - merr.Add(c.Close()) - } - return merr.Err() -} diff --git a/tsdb/chunks/head_chunks.go b/tsdb/chunks/head_chunks.go index 6f67b8318..db45280fd 100644 --- a/tsdb/chunks/head_chunks.go +++ b/tsdb/chunks/head_chunks.go @@ -153,10 +153,7 @@ func (cdm *ChunkDiskMapper) openMMapFiles() (returnErr error) { cdm.closers = map[int]io.Closer{} defer func() { if returnErr != nil { - var merr tsdb_errors.MultiError - merr.Add(returnErr) - merr.Add(closeAllFromMap(cdm.closers)) - returnErr = merr.Err() + returnErr = tsdb_errors.NewMulti(returnErr, closeAllFromMap(cdm.closers)).Err() cdm.mmappedChunkFiles = nil cdm.closers = nil @@ -370,10 +367,7 @@ func (cdm *ChunkDiskMapper) cut() (returnErr error) { // The file should not be closed if there is no error, // its kept open in the ChunkDiskMapper. if returnErr != nil { - var merr tsdb_errors.MultiError - merr.Add(returnErr) - merr.Add(newFile.Close()) - returnErr = merr.Err() + returnErr = tsdb_errors.NewMulti(returnErr, newFile.Close()).Err() } }() @@ -717,13 +711,13 @@ func (cdm *ChunkDiskMapper) Truncate(mint int64) error { } cdm.readPathMtx.RUnlock() - var merr tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() // Cut a new file only if the current file has some chunks. if cdm.curFileSize() > HeadChunkFileHeaderSize { - merr.Add(cdm.CutNewFile()) + errs.Add(cdm.CutNewFile()) } - merr.Add(cdm.deleteFiles(removedFiles)) - return merr.Err() + errs.Add(cdm.deleteFiles(removedFiles)) + return errs.Err() } func (cdm *ChunkDiskMapper) deleteFiles(removedFiles []int) error { @@ -795,23 +789,23 @@ func (cdm *ChunkDiskMapper) Close() error { } cdm.closed = true - var merr tsdb_errors.MultiError - merr.Add(closeAllFromMap(cdm.closers)) - merr.Add(cdm.finalizeCurFile()) - merr.Add(cdm.dir.Close()) - + errs := tsdb_errors.NewMulti( + closeAllFromMap(cdm.closers), + cdm.finalizeCurFile(), + cdm.dir.Close(), + ) cdm.mmappedChunkFiles = map[int]*mmappedChunkFile{} cdm.closers = map[int]io.Closer{} - return merr.Err() + return errs.Err() } func closeAllFromMap(cs map[int]io.Closer) error { - var merr tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for _, c := range cs { - merr.Add(c.Close()) + errs.Add(c.Close()) } - return merr.Err() + return errs.Err() } const inBufferShards = 128 // 128 is a randomly chosen number. diff --git a/tsdb/compact.go b/tsdb/compact.go index cd70ee52d..9026b37bf 100644 --- a/tsdb/compact.go +++ b/tsdb/compact.go @@ -451,17 +451,16 @@ func (c *LeveledCompactor) Compact(dest string, dirs []string, open []*Block) (u return uid, nil } - var merr tsdb_errors.MultiError - merr.Add(err) + errs := tsdb_errors.NewMulti(err) if err != context.Canceled { for _, b := range bs { if err := b.setCompactionFailed(); err != nil { - merr.Add(errors.Wrapf(err, "setting compaction failed for block: %s", b.Dir())) + errs.Add(errors.Wrapf(err, "setting compaction failed for block: %s", b.Dir())) } } } - return uid, merr + return uid, errs.Err() } func (c *LeveledCompactor) Write(dest string, b BlockReader, mint, maxt int64, parent *BlockMeta) (ulid.ULID, error) { @@ -534,10 +533,7 @@ func (c *LeveledCompactor) write(dest string, meta *BlockMeta, blocks ...BlockRe tmp := dir + tmpForCreationBlockDirSuffix var closers []io.Closer defer func(t time.Time) { - var merr tsdb_errors.MultiError - merr.Add(err) - merr.Add(closeAll(closers)) - err = merr.Err() + err = tsdb_errors.NewMulti(err, tsdb_errors.CloseAll(closers)).Err() // RemoveAll returns no error when tmp doesn't exist so it is safe to always run it. if err := os.RemoveAll(tmp); err != nil { @@ -594,13 +590,13 @@ func (c *LeveledCompactor) write(dest string, meta *BlockMeta, blocks ...BlockRe // though these are covered under defer. This is because in Windows, // you cannot delete these unless they are closed and the defer is to // make sure they are closed if the function exits due to an error above. - var merr tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for _, w := range closers { - merr.Add(w.Close()) + errs.Add(w.Close()) } closers = closers[:0] // Avoid closing the writers twice in the defer. - if merr.Err() != nil { - return merr.Err() + if errs.Err() != nil { + return errs.Err() } // Populated block is empty, so exit early. @@ -660,12 +656,11 @@ func (c *LeveledCompactor) populateBlock(blocks []BlockReader, meta *BlockMeta, overlapping bool ) defer func() { - var merr tsdb_errors.MultiError - merr.Add(err) - if cerr := closeAll(closers); cerr != nil { - merr.Add(errors.Wrap(cerr, "close")) + errs := tsdb_errors.NewMulti(err) + if cerr := tsdb_errors.CloseAll(closers); cerr != nil { + errs.Add(errors.Wrap(cerr, "close")) } - err = merr.Err() + err = errs.Err() c.metrics.populatingBlocks.Set(0) }() c.metrics.populatingBlocks.Set(1) diff --git a/tsdb/db.go b/tsdb/db.go index 4c67e08fa..81669829d 100644 --- a/tsdb/db.go +++ b/tsdb/db.go @@ -333,10 +333,10 @@ func (db *DBReadOnly) FlushWAL(dir string) (returnErr error) { return err } defer func() { - var merr tsdb_errors.MultiError - merr.Add(returnErr) - merr.Add(errors.Wrap(head.Close(), "closing Head")) - returnErr = merr.Err() + returnErr = tsdb_errors.NewMulti( + returnErr, + errors.Wrap(head.Close(), "closing Head"), + ).Err() }() // Set the min valid time for the ingested wal samples // to be no lower than the maxt of the last block. @@ -466,11 +466,11 @@ func (db *DBReadOnly) Blocks() ([]BlockReader, error) { level.Warn(db.logger).Log("msg", "Closing block failed", "err", err, "block", b) } } - var merr tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for ulid, err := range corrupted { - merr.Add(errors.Wrapf(err, "corrupted block %s", ulid.String())) + errs.Add(errors.Wrapf(err, "corrupted block %s", ulid.String())) } - return nil, merr.Err() + return nil, errs.Err() } if len(loadable) == 0 { @@ -514,12 +514,7 @@ func (db *DBReadOnly) Close() error { } close(db.closed) - var merr tsdb_errors.MultiError - - for _, b := range db.closers { - merr.Add(b.Close()) - } - return merr.Err() + return tsdb_errors.CloseAll(db.closers) } // Open returns a new DB in the given directory. If options are empty, DefaultOptions will be used. @@ -602,11 +597,10 @@ func open(dir string, l log.Logger, r prometheus.Registerer, opts *Options, rngs close(db.donec) // DB is never run if it was an error, so close this channel here. - var merr tsdb_errors.MultiError - merr.Add(returnedErr) - merr.Add(errors.Wrap(db.Close(), "close DB after failed startup")) - - returnedErr = merr.Err() + returnedErr = tsdb_errors.NewMulti( + returnedErr, + errors.Wrap(db.Close(), "close DB after failed startup"), + ).Err() }() if db.blocksToDelete == nil { @@ -798,10 +792,10 @@ func (db *DB) Compact() (returnErr error) { lastBlockMaxt := int64(math.MinInt64) defer func() { - var merr tsdb_errors.MultiError - merr.Add(returnErr) - merr.Add(errors.Wrap(db.head.truncateWAL(lastBlockMaxt), "WAL truncation in Compact defer")) - returnErr = merr.Err() + returnErr = tsdb_errors.NewMulti( + returnErr, + errors.Wrap(db.head.truncateWAL(lastBlockMaxt), "WAL truncation in Compact defer"), + ).Err() }() // Check whether we have pending head blocks that are ready to be persisted. @@ -867,10 +861,10 @@ func (db *DB) compactHead(head *RangeHead) error { runtime.GC() if err := db.reloadBlocks(); err != nil { if errRemoveAll := os.RemoveAll(filepath.Join(db.dir, uid.String())); errRemoveAll != nil { - var merr tsdb_errors.MultiError - merr.Add(errors.Wrap(err, "reloadBlocks blocks")) - merr.Add(errors.Wrapf(errRemoveAll, "delete persisted head block after failed db reloadBlocks:%s", uid)) - return merr.Err() + return tsdb_errors.NewMulti( + errors.Wrap(err, "reloadBlocks blocks"), + errors.Wrapf(errRemoveAll, "delete persisted head block after failed db reloadBlocks:%s", uid), + ).Err() } return errors.Wrap(err, "reloadBlocks blocks") } @@ -984,11 +978,11 @@ func (db *DB) reloadBlocks() (err error) { block.Close() } } - var merr tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for ulid, err := range corrupted { - merr.Add(errors.Wrapf(err, "corrupted block %s", ulid.String())) + errs.Add(errors.Wrapf(err, "corrupted block %s", ulid.String())) } - return merr.Err() + return errs.Err() } var ( @@ -1343,17 +1337,14 @@ func (db *DB) Close() error { g.Go(pb.Close) } - var merr tsdb_errors.MultiError - - merr.Add(g.Wait()) - + errs := tsdb_errors.NewMulti(g.Wait()) if db.lockf != nil { - merr.Add(db.lockf.Release()) + errs.Add(db.lockf.Release()) } if db.head != nil { - merr.Add(db.head.Close()) + errs.Add(db.head.Close()) } - return merr.Err() + return errs.Err() } // DisableCompactions disables auto compactions. @@ -1618,15 +1609,6 @@ func nextSequenceFile(dir string) (string, int, error) { return filepath.Join(dir, fmt.Sprintf("%0.6d", i+1)), int(i + 1), nil } -func closeAll(cs []io.Closer) error { - var merr tsdb_errors.MultiError - - for _, c := range cs { - merr.Add(c.Close()) - } - return merr.Err() -} - func exponential(d, min, max time.Duration) time.Duration { d *= 2 if d < min { diff --git a/tsdb/errors/errors.go b/tsdb/errors/errors.go index 69d366248..aeac4d277 100644 --- a/tsdb/errors/errors.go +++ b/tsdb/errors/errors.go @@ -17,21 +17,59 @@ package errors import ( "bytes" "fmt" + "io" ) -// The MultiError type implements the error interface, and contains the -// Errors used to construct it. -type MultiError []error +// multiError type allows combining multiple errors into one. +type multiError []error -// Returns a concatenated string of the contained errors -func (es MultiError) Error() string { +// NewMulti returns multiError with provided errors added if not nil. +func NewMulti(errs ...error) multiError { // nolint:golint + m := multiError{} + m.Add(errs...) + return m +} + +// Add adds single or many errors to the error list. Each error is added only if not nil. +// If the error is a nonNilMultiError type, the errors inside nonNilMultiError are added to the main multiError. +func (es *multiError) Add(errs ...error) { + for _, err := range errs { + if err == nil { + continue + } + if merr, ok := err.(nonNilMultiError); ok { + *es = append(*es, merr.errs...) + continue + } + *es = append(*es, err) + } +} + +// Err returns the error list as an error or nil if it is empty. +func (es multiError) Err() error { + if len(es) == 0 { + return nil + } + return nonNilMultiError{errs: es} +} + +// nonNilMultiError implements the error interface, and it represents +// multiError with at least one error inside it. +// This type is needed to make sure that nil is returned when no error is combined in multiError for err != nil +// check to work. +type nonNilMultiError struct { + errs multiError +} + +// Error returns a concatenated string of the contained errors. +func (es nonNilMultiError) Error() string { var buf bytes.Buffer - if len(es) > 1 { - fmt.Fprintf(&buf, "%d errors: ", len(es)) + if len(es.errs) > 1 { + fmt.Fprintf(&buf, "%d errors: ", len(es.errs)) } - for i, err := range es { + for i, err := range es.errs { if i != 0 { buf.WriteString("; ") } @@ -41,22 +79,11 @@ func (es MultiError) Error() string { return buf.String() } -// Add adds the error to the error list if it is not nil. -func (es *MultiError) Add(err error) { - if err == nil { - return - } - if merr, ok := err.(MultiError); ok { - *es = append(*es, merr...) - } else { - *es = append(*es, err) +// CloseAll closes all given closers while recording error in MultiError. +func CloseAll(cs []io.Closer) error { + errs := NewMulti() + for _, c := range cs { + errs.Add(c.Close()) } -} - -// Err returns the error list as an error or nil if it is empty. -func (es MultiError) Err() error { - if len(es) == 0 { - return nil - } - return es + return errs.Err() } diff --git a/tsdb/head.go b/tsdb/head.go index e4dfef20a..dad122778 100644 --- a/tsdb/head.go +++ b/tsdb/head.go @@ -1465,12 +1465,11 @@ func (h *Head) Close() error { h.closedMtx.Lock() defer h.closedMtx.Unlock() h.closed = true - var merr tsdb_errors.MultiError - merr.Add(h.chunkDiskMapper.Close()) + errs := tsdb_errors.NewMulti(h.chunkDiskMapper.Close()) if h.wal != nil { - merr.Add(h.wal.Close()) + errs.Add(h.wal.Close()) } - return merr.Err() + return errs.Err() } type headChunkReader struct { diff --git a/tsdb/index/index.go b/tsdb/index/index.go index b3d8293f1..c747f5eee 100644 --- a/tsdb/index/index.go +++ b/tsdb/index/index.go @@ -1075,10 +1075,10 @@ func NewFileReader(path string) (*Reader, error) { } r, err := newReader(realByteSlice(f.Bytes()), f) if err != nil { - var merr tsdb_errors.MultiError - merr.Add(err) - merr.Add(f.Close()) - return nil, merr + return nil, tsdb_errors.NewMulti( + err, + f.Close(), + ).Err() } return r, nil diff --git a/tsdb/querier.go b/tsdb/querier.go index 63c311ee2..f666bb790 100644 --- a/tsdb/querier.go +++ b/tsdb/querier.go @@ -97,12 +97,14 @@ func (q *blockBaseQuerier) Close() error { if q.closed { return errors.New("block querier already closed") } - var merr tsdb_errors.MultiError - merr.Add(q.index.Close()) - merr.Add(q.chunks.Close()) - merr.Add(q.tombstones.Close()) + + errs := tsdb_errors.NewMulti( + q.index.Close(), + q.chunks.Close(), + q.tombstones.Close(), + ) q.closed = true - return merr.Err() + return errs.Err() } type blockQuerier struct { diff --git a/tsdb/repair.go b/tsdb/repair.go index 49f55d5a8..cc777546e 100644 --- a/tsdb/repair.go +++ b/tsdb/repair.go @@ -83,18 +83,18 @@ func repairBadIndexVersion(logger log.Logger, dir string) error { return errors.Wrapf(err, "copy content of index to index.repaired for block dir: %v", d) } - var merr tsdb_errors.MultiError - // Set the 5th byte to 2 to indicate the correct file format version. if _, err := repl.WriteAt([]byte{2}, 4); err != nil { - merr.Add(errors.Wrap(err, "rewrite of index.repaired")) - merr.Add(errors.Wrap(repl.Close(), "close")) - return errors.Wrapf(merr.Err(), "block dir: %v", d) + return tsdb_errors.NewMulti( + errors.Wrapf(err, "rewrite of index.repaired for block dir: %v", d), + errors.Wrap(repl.Close(), "close"), + ).Err() } if err := repl.Sync(); err != nil { - merr.Add(errors.Wrap(err, "sync of index.repaired")) - merr.Add(errors.Wrap(repl.Close(), "close")) - return errors.Wrapf(merr.Err(), "block dir: %v", d) + return tsdb_errors.NewMulti( + errors.Wrapf(err, "sync of index.repaired for block dir: %v", d), + errors.Wrap(repl.Close(), "close"), + ).Err() } if err := repl.Close(); err != nil { return errors.Wrapf(repl.Close(), "close repaired index for block dir: %v", d) diff --git a/tsdb/tombstones/tombstones.go b/tsdb/tombstones/tombstones.go index 79b68adfb..aa74a7d1e 100644 --- a/tsdb/tombstones/tombstones.go +++ b/tsdb/tombstones/tombstones.go @@ -126,10 +126,8 @@ func WriteFile(logger log.Logger, dir string, tr Reader) (int64, error) { } size += n - var merr tsdb_errors.MultiError - if merr.Add(f.Sync()); merr.Err() != nil { - merr.Add(f.Close()) - return 0, merr.Err() + if err := f.Sync(); err != nil { + return 0, tsdb_errors.NewMulti(err, f.Close()).Err() } if err = f.Close(); err != nil { diff --git a/tsdb/wal/checkpoint.go b/tsdb/wal/checkpoint.go index c27f6f5b9..a264e1e95 100644 --- a/tsdb/wal/checkpoint.go +++ b/tsdb/wal/checkpoint.go @@ -68,14 +68,12 @@ func DeleteCheckpoints(dir string, maxIndex int) error { return err } - var errs tsdb_errors.MultiError + errs := tsdb_errors.NewMulti() for _, checkpoint := range checkpoints { if checkpoint.index >= maxIndex { break } - if err := os.RemoveAll(filepath.Join(dir, checkpoint.name)); err != nil { - errs.Add(err) - } + errs.Add(os.RemoveAll(filepath.Join(dir, checkpoint.name))) } return errs.Err() } diff --git a/tsdb/wal/reader_test.go b/tsdb/wal/reader_test.go index 340d9a3be..b0b9a12c0 100644 --- a/tsdb/wal/reader_test.go +++ b/tsdb/wal/reader_test.go @@ -280,11 +280,7 @@ func (m *multiReadCloser) Read(p []byte) (n int, err error) { return m.reader.Read(p) } func (m *multiReadCloser) Close() error { - var merr tsdb_errors.MultiError - for _, closer := range m.closers { - merr.Add(closer.Close()) - } - return merr.Err() + return tsdb_errors.NewMulti(tsdb_errors.CloseAll(m.closers)).Err() } func allSegments(dir string) (io.ReadCloser, error) {