mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-12-30 07:40:06 +01:00
384 lines
13 KiB
Go
384 lines
13 KiB
Go
|
package logstorage
|
||
|
|
||
|
import (
|
||
|
"path/filepath"
|
||
|
"sync"
|
||
|
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/filestream"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fs"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
|
||
|
)
|
||
|
|
||
|
type readerWithStats struct {
|
||
|
r filestream.ReadCloser
|
||
|
bytesRead uint64
|
||
|
}
|
||
|
|
||
|
func (r *readerWithStats) reset() {
|
||
|
r.r = nil
|
||
|
r.bytesRead = 0
|
||
|
}
|
||
|
|
||
|
func (r *readerWithStats) init(rc filestream.ReadCloser) {
|
||
|
r.reset()
|
||
|
|
||
|
r.r = rc
|
||
|
}
|
||
|
|
||
|
// Path returns the path to r file
|
||
|
func (r *readerWithStats) Path() string {
|
||
|
return r.r.Path()
|
||
|
}
|
||
|
|
||
|
// MustReadFull reads len(data) to r.
|
||
|
func (r *readerWithStats) MustReadFull(data []byte) {
|
||
|
fs.MustReadData(r.r, data)
|
||
|
r.bytesRead += uint64(len(data))
|
||
|
}
|
||
|
|
||
|
func (r *readerWithStats) Read(p []byte) (int, error) {
|
||
|
n, err := r.r.Read(p)
|
||
|
r.bytesRead += uint64(n)
|
||
|
return n, err
|
||
|
}
|
||
|
|
||
|
func (r *readerWithStats) MustClose() {
|
||
|
r.r.MustClose()
|
||
|
r.r = nil
|
||
|
}
|
||
|
|
||
|
// streamReaders contains readers for blockStreamReader
|
||
|
type streamReaders struct {
|
||
|
metaindexReader readerWithStats
|
||
|
indexReader readerWithStats
|
||
|
columnsHeaderReader readerWithStats
|
||
|
timestampsReader readerWithStats
|
||
|
fieldValuesReader readerWithStats
|
||
|
fieldBloomFilterReader readerWithStats
|
||
|
messageValuesReader readerWithStats
|
||
|
messageBloomFilterReader readerWithStats
|
||
|
}
|
||
|
|
||
|
func (sr *streamReaders) reset() {
|
||
|
sr.metaindexReader.reset()
|
||
|
sr.indexReader.reset()
|
||
|
sr.columnsHeaderReader.reset()
|
||
|
sr.timestampsReader.reset()
|
||
|
sr.fieldValuesReader.reset()
|
||
|
sr.fieldBloomFilterReader.reset()
|
||
|
sr.messageValuesReader.reset()
|
||
|
sr.messageBloomFilterReader.reset()
|
||
|
}
|
||
|
|
||
|
func (sr *streamReaders) init(metaindexReader, indexReader, columnsHeaderReader, timestampsReader, fieldValuesReader, fieldBloomFilterReader,
|
||
|
messageValuesReader, messageBloomFilterReader filestream.ReadCloser,
|
||
|
) {
|
||
|
sr.metaindexReader.init(metaindexReader)
|
||
|
sr.indexReader.init(indexReader)
|
||
|
sr.columnsHeaderReader.init(columnsHeaderReader)
|
||
|
sr.timestampsReader.init(timestampsReader)
|
||
|
sr.fieldValuesReader.init(fieldValuesReader)
|
||
|
sr.fieldBloomFilterReader.init(fieldBloomFilterReader)
|
||
|
sr.messageValuesReader.init(messageValuesReader)
|
||
|
sr.messageBloomFilterReader.init(messageBloomFilterReader)
|
||
|
}
|
||
|
|
||
|
func (sr *streamReaders) totalBytesRead() uint64 {
|
||
|
n := uint64(0)
|
||
|
n += sr.metaindexReader.bytesRead
|
||
|
n += sr.indexReader.bytesRead
|
||
|
n += sr.columnsHeaderReader.bytesRead
|
||
|
n += sr.timestampsReader.bytesRead
|
||
|
n += sr.fieldValuesReader.bytesRead
|
||
|
n += sr.fieldBloomFilterReader.bytesRead
|
||
|
n += sr.messageValuesReader.bytesRead
|
||
|
n += sr.messageBloomFilterReader.bytesRead
|
||
|
return n
|
||
|
}
|
||
|
|
||
|
func (sr *streamReaders) MustClose() {
|
||
|
sr.metaindexReader.MustClose()
|
||
|
sr.indexReader.MustClose()
|
||
|
sr.columnsHeaderReader.MustClose()
|
||
|
sr.timestampsReader.MustClose()
|
||
|
sr.fieldValuesReader.MustClose()
|
||
|
sr.fieldBloomFilterReader.MustClose()
|
||
|
sr.messageValuesReader.MustClose()
|
||
|
sr.messageBloomFilterReader.MustClose()
|
||
|
}
|
||
|
|
||
|
// blockStreamReader is used for reading blocks in streaming manner from a part.
|
||
|
type blockStreamReader struct {
|
||
|
// blockData contains the data for the last read block
|
||
|
blockData blockData
|
||
|
|
||
|
// ph is the header for the part
|
||
|
ph partHeader
|
||
|
|
||
|
// streamReaders contains data readers in stream mode
|
||
|
streamReaders streamReaders
|
||
|
|
||
|
// indexBlockHeaders contains the list of all the indexBlockHeader entries for the part
|
||
|
indexBlockHeaders []indexBlockHeader
|
||
|
|
||
|
// blockHeaders contains the list of blockHeader entries for the current indexBlockHeader pointed by nextIndexBlockIdx
|
||
|
blockHeaders []blockHeader
|
||
|
|
||
|
// nextIndexBlockIdx is the index of the next item to read from indexBlockHeaders
|
||
|
nextIndexBlockIdx int
|
||
|
|
||
|
// nextBlockIdx is the index of the next item to read from blockHeaders
|
||
|
nextBlockIdx int
|
||
|
|
||
|
// globalUncompressedSizeBytes is the total size of log entries seen in the part
|
||
|
globalUncompressedSizeBytes uint64
|
||
|
|
||
|
// globalRowsCount is the number of log entries seen in the part
|
||
|
globalRowsCount uint64
|
||
|
|
||
|
// globalBlocksCount is the number of blocks seen in the part
|
||
|
globalBlocksCount uint64
|
||
|
|
||
|
// sidLast is the stream id for the previously read block
|
||
|
sidLast streamID
|
||
|
|
||
|
// minTimestampLast is the minimum timestamp for the previously read block
|
||
|
minTimestampLast int64
|
||
|
}
|
||
|
|
||
|
// reset resets bsr, so it can be re-used
|
||
|
func (bsr *blockStreamReader) reset() {
|
||
|
bsr.blockData.reset()
|
||
|
bsr.ph.reset()
|
||
|
bsr.streamReaders.reset()
|
||
|
|
||
|
ihs := bsr.indexBlockHeaders
|
||
|
if len(ihs) > 10e3 {
|
||
|
// The ihs len is unbound, so it is better to drop too long indexBlockHeaders in order to reduce memory usage
|
||
|
ihs = nil
|
||
|
}
|
||
|
for i := range ihs {
|
||
|
ihs[i].reset()
|
||
|
}
|
||
|
bsr.indexBlockHeaders = ihs[:0]
|
||
|
|
||
|
bhs := bsr.blockHeaders
|
||
|
for i := range bhs {
|
||
|
bhs[i].reset()
|
||
|
}
|
||
|
bsr.blockHeaders = bhs[:0]
|
||
|
|
||
|
bsr.nextIndexBlockIdx = 0
|
||
|
bsr.nextBlockIdx = 0
|
||
|
bsr.globalUncompressedSizeBytes = 0
|
||
|
bsr.globalRowsCount = 0
|
||
|
bsr.globalBlocksCount = 0
|
||
|
|
||
|
bsr.sidLast.reset()
|
||
|
bsr.minTimestampLast = 0
|
||
|
}
|
||
|
|
||
|
// Path returns part path for bsr (e.g. file path, url or in-memory reference)
|
||
|
func (bsr *blockStreamReader) Path() string {
|
||
|
path := bsr.streamReaders.metaindexReader.Path()
|
||
|
return filepath.Dir(path)
|
||
|
}
|
||
|
|
||
|
// MustInitFromInmemoryPart initializes bsr from mp.
|
||
|
func (bsr *blockStreamReader) MustInitFromInmemoryPart(mp *inmemoryPart) {
|
||
|
bsr.reset()
|
||
|
|
||
|
bsr.ph = mp.ph
|
||
|
|
||
|
// Initialize streamReaders
|
||
|
metaindexReader := mp.metaindex.NewReader()
|
||
|
indexReader := mp.index.NewReader()
|
||
|
columnsHeaderReader := mp.columnsHeader.NewReader()
|
||
|
timestampsReader := mp.timestamps.NewReader()
|
||
|
fieldValuesReader := mp.fieldValues.NewReader()
|
||
|
fieldBloomFilterReader := mp.fieldBloomFilter.NewReader()
|
||
|
messageValuesReader := mp.messageValues.NewReader()
|
||
|
messageBloomFilterReader := mp.messageBloomFilter.NewReader()
|
||
|
|
||
|
bsr.streamReaders.init(metaindexReader, indexReader, columnsHeaderReader, timestampsReader,
|
||
|
fieldValuesReader, fieldBloomFilterReader, messageValuesReader, messageBloomFilterReader)
|
||
|
|
||
|
// Read metaindex data
|
||
|
bsr.indexBlockHeaders = mustReadIndexBlockHeaders(bsr.indexBlockHeaders[:0], &bsr.streamReaders.metaindexReader)
|
||
|
}
|
||
|
|
||
|
// MustInitFromFilePart initializes bsr from file part at the given path.
|
||
|
func (bsr *blockStreamReader) MustInitFromFilePart(path string) {
|
||
|
bsr.reset()
|
||
|
|
||
|
// Files in the part are always read without OS cache pollution,
|
||
|
// since they are usually deleted after the merge.
|
||
|
const nocache = true
|
||
|
|
||
|
metaindexPath := filepath.Join(path, metaindexFilename)
|
||
|
indexPath := filepath.Join(path, indexFilename)
|
||
|
columnsHeaderPath := filepath.Join(path, columnsHeaderFilename)
|
||
|
timestampsPath := filepath.Join(path, timestampsFilename)
|
||
|
fieldValuesPath := filepath.Join(path, fieldValuesFilename)
|
||
|
fieldBloomFilterPath := filepath.Join(path, fieldBloomFilename)
|
||
|
messageValuesPath := filepath.Join(path, messageValuesFilename)
|
||
|
messageBloomFilterPath := filepath.Join(path, messageBloomFilename)
|
||
|
|
||
|
bsr.ph.mustReadMetadata(path)
|
||
|
|
||
|
// Open data readers
|
||
|
metaindexReader := filestream.MustOpen(metaindexPath, nocache)
|
||
|
indexReader := filestream.MustOpen(indexPath, nocache)
|
||
|
columnsHeaderReader := filestream.MustOpen(columnsHeaderPath, nocache)
|
||
|
timestampsReader := filestream.MustOpen(timestampsPath, nocache)
|
||
|
fieldValuesReader := filestream.MustOpen(fieldValuesPath, nocache)
|
||
|
fieldBloomFilterReader := filestream.MustOpen(fieldBloomFilterPath, nocache)
|
||
|
messageValuesReader := filestream.MustOpen(messageValuesPath, nocache)
|
||
|
messageBloomFilterReader := filestream.MustOpen(messageBloomFilterPath, nocache)
|
||
|
|
||
|
// Initialize streamReaders
|
||
|
bsr.streamReaders.init(metaindexReader, indexReader, columnsHeaderReader, timestampsReader,
|
||
|
fieldValuesReader, fieldBloomFilterReader, messageValuesReader, messageBloomFilterReader)
|
||
|
|
||
|
// Read metaindex data
|
||
|
bsr.indexBlockHeaders = mustReadIndexBlockHeaders(bsr.indexBlockHeaders[:0], &bsr.streamReaders.metaindexReader)
|
||
|
}
|
||
|
|
||
|
// NextBlock reads the next block from bsr and puts it into bsr.blockData.
|
||
|
//
|
||
|
// false is returned if there are no other blocks.
|
||
|
func (bsr *blockStreamReader) NextBlock() bool {
|
||
|
for bsr.nextBlockIdx >= len(bsr.blockHeaders) {
|
||
|
if !bsr.nextIndexBlock() {
|
||
|
return false
|
||
|
}
|
||
|
}
|
||
|
ih := &bsr.indexBlockHeaders[bsr.nextIndexBlockIdx-1]
|
||
|
bh := &bsr.blockHeaders[bsr.nextBlockIdx]
|
||
|
th := &bh.timestampsHeader
|
||
|
|
||
|
// Validate bh
|
||
|
if bh.streamID.less(&bsr.sidLast) {
|
||
|
logger.Panicf("FATAL: %s: blockHeader.streamID=%s cannot be smaller than the streamID from the previously read block: %s", bsr.Path(), &bh.streamID, &bsr.sidLast)
|
||
|
}
|
||
|
if bh.streamID.equal(&bsr.sidLast) && th.minTimestamp < bsr.minTimestampLast {
|
||
|
logger.Panicf("FATAL: %s: timestamps.minTimestamp=%d cannot be smaller than the minTimestamp for the previously read block for the same streamID: %d",
|
||
|
bsr.Path(), th.minTimestamp, bsr.minTimestampLast)
|
||
|
}
|
||
|
bsr.minTimestampLast = th.minTimestamp
|
||
|
bsr.sidLast = bh.streamID
|
||
|
if th.minTimestamp < ih.minTimestamp {
|
||
|
logger.Panicf("FATAL: %s: timestampsHeader.minTimestamp=%d cannot be smaller than indexBlockHeader.minTimestamp=%d", bsr.Path(), th.minTimestamp, ih.minTimestamp)
|
||
|
}
|
||
|
if th.maxTimestamp > ih.maxTimestamp {
|
||
|
logger.Panicf("FATAL: %s: timestampsHeader.maxTimestamp=%d cannot be bigger than indexBlockHeader.maxTimestamp=%d", bsr.Path(), th.maxTimestamp, ih.minTimestamp)
|
||
|
}
|
||
|
|
||
|
// Read bsr.blockData
|
||
|
bsr.blockData.mustReadFrom(bh, &bsr.streamReaders)
|
||
|
|
||
|
bsr.globalUncompressedSizeBytes += bh.uncompressedSizeBytes
|
||
|
bsr.globalRowsCount += bh.rowsCount
|
||
|
bsr.globalBlocksCount++
|
||
|
if bsr.globalUncompressedSizeBytes > bsr.ph.UncompressedSizeBytes {
|
||
|
logger.Panicf("FATAL: %s: too big size of entries read: %d; mustn't exceed partHeader.UncompressedSizeBytes=%d",
|
||
|
bsr.Path(), bsr.globalUncompressedSizeBytes, bsr.ph.UncompressedSizeBytes)
|
||
|
}
|
||
|
if bsr.globalRowsCount > bsr.ph.RowsCount {
|
||
|
logger.Panicf("FATAL: %s: too many log entries read so far: %d; mustn't exceed partHeader.RowsCount=%d", bsr.Path(), bsr.globalRowsCount, bsr.ph.RowsCount)
|
||
|
}
|
||
|
if bsr.globalBlocksCount > bsr.ph.BlocksCount {
|
||
|
logger.Panicf("FATAL: %s: too many blocks read so far: %d; mustn't exceed partHeader.BlocksCount=%d", bsr.Path(), bsr.globalBlocksCount, bsr.ph.BlocksCount)
|
||
|
}
|
||
|
|
||
|
// The block has been sucessfully read
|
||
|
bsr.nextBlockIdx++
|
||
|
return true
|
||
|
}
|
||
|
|
||
|
func (bsr *blockStreamReader) nextIndexBlock() bool {
|
||
|
// Advance to the next indexBlockHeader
|
||
|
if bsr.nextIndexBlockIdx >= len(bsr.indexBlockHeaders) {
|
||
|
// No more blocks left
|
||
|
// Validate bsr.ph
|
||
|
totalBytesRead := bsr.streamReaders.totalBytesRead()
|
||
|
if bsr.ph.CompressedSizeBytes != totalBytesRead {
|
||
|
logger.Panicf("FATAL: %s: partHeader.CompressedSizeBytes=%d must match the size of data read: %d", bsr.Path(), bsr.ph.CompressedSizeBytes, totalBytesRead)
|
||
|
}
|
||
|
if bsr.ph.UncompressedSizeBytes != bsr.globalUncompressedSizeBytes {
|
||
|
logger.Panicf("FATAL: %s: partHeader.UncompressedSizeBytes=%d must match the size of entries read: %d",
|
||
|
bsr.Path(), bsr.ph.UncompressedSizeBytes, bsr.globalUncompressedSizeBytes)
|
||
|
}
|
||
|
if bsr.ph.RowsCount != bsr.globalRowsCount {
|
||
|
logger.Panicf("FATAL: %s: partHeader.RowsCount=%d must match the number of log entries read: %d", bsr.Path(), bsr.ph.RowsCount, bsr.globalRowsCount)
|
||
|
}
|
||
|
if bsr.ph.BlocksCount != bsr.globalBlocksCount {
|
||
|
logger.Panicf("FATAL: %s: partHeader.BlocksCount=%d must match the number of blocks read: %d", bsr.Path(), bsr.ph.BlocksCount, bsr.globalBlocksCount)
|
||
|
}
|
||
|
return false
|
||
|
}
|
||
|
ih := &bsr.indexBlockHeaders[bsr.nextIndexBlockIdx]
|
||
|
|
||
|
// Validate ih
|
||
|
metaindexReader := &bsr.streamReaders.metaindexReader
|
||
|
if ih.minTimestamp < bsr.ph.MinTimestamp {
|
||
|
logger.Panicf("FATAL: %s: indexBlockHeader.minTimestamp=%d cannot be smaller than partHeader.MinTimestamp=%d",
|
||
|
metaindexReader.Path(), ih.minTimestamp, bsr.ph.MinTimestamp)
|
||
|
}
|
||
|
if ih.maxTimestamp > bsr.ph.MaxTimestamp {
|
||
|
logger.Panicf("FATAL: %s: indexBlockHeader.maxTimestamp=%d cannot be bigger than partHeader.MaxTimestamp=%d",
|
||
|
metaindexReader.Path(), ih.maxTimestamp, bsr.ph.MaxTimestamp)
|
||
|
}
|
||
|
|
||
|
// Read indexBlock for the given ih
|
||
|
bb := longTermBufPool.Get()
|
||
|
bb.B = ih.mustReadNextIndexBlock(bb.B[:0], &bsr.streamReaders)
|
||
|
bsr.blockHeaders = resetBlockHeaders(bsr.blockHeaders)
|
||
|
var err error
|
||
|
bsr.blockHeaders, err = unmarshalBlockHeaders(bsr.blockHeaders[:0], bb.B)
|
||
|
longTermBufPool.Put(bb)
|
||
|
if err != nil {
|
||
|
logger.Panicf("FATAL: %s: cannot unmarshal blockHeader entries: %s", bsr.streamReaders.indexReader.Path(), err)
|
||
|
}
|
||
|
|
||
|
bsr.nextIndexBlockIdx++
|
||
|
bsr.nextBlockIdx = 0
|
||
|
return true
|
||
|
}
|
||
|
|
||
|
// MustClose closes bsr.
|
||
|
func (bsr *blockStreamReader) MustClose() {
|
||
|
bsr.streamReaders.MustClose()
|
||
|
bsr.reset()
|
||
|
}
|
||
|
|
||
|
// getBlockStreamReader returns blockStreamReader.
|
||
|
//
|
||
|
// The returned blockStreamReader must be initialized with MustInit().
|
||
|
// call putBlockStreamReader() when the retruend blockStreamReader is no longer needed.
|
||
|
func getBlockStreamReader() *blockStreamReader {
|
||
|
v := blockStreamReaderPool.Get()
|
||
|
if v == nil {
|
||
|
v = &blockStreamReader{}
|
||
|
}
|
||
|
bsr := v.(*blockStreamReader)
|
||
|
return bsr
|
||
|
}
|
||
|
|
||
|
// putBlockStreamReader returns bsr to the pool.
|
||
|
//
|
||
|
// bsr cannot be used after returning to the pool.
|
||
|
func putBlockStreamReader(bsr *blockStreamReader) {
|
||
|
bsr.reset()
|
||
|
blockStreamReaderPool.Put(bsr)
|
||
|
}
|
||
|
|
||
|
var blockStreamReaderPool sync.Pool
|
||
|
|
||
|
// mustCloseBlockStreamReaders calls MustClose() on the given bsrs.
|
||
|
func mustCloseBlockStreamReaders(bsrs []*blockStreamReader) {
|
||
|
for _, bsr := range bsrs {
|
||
|
bsr.MustClose()
|
||
|
}
|
||
|
}
|