mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-12-15 16:30:55 +01:00
29bbab0ec9
The prioritizing could lead to big merge starvation, which could end up in too big number of parts that must be merged into big parts. Multiple big merges may be initiated after the migration from v1.39.0 or v1.39.1. It is OK - these merges should be finished soon, which should return CPU and disk IO usage to normal levels. Updates https://github.com/VictoriaMetrics/VictoriaMetrics/issues/648 Updates https://github.com/VictoriaMetrics/VictoriaMetrics/issues/618
116 lines
3.1 KiB
Go
116 lines
3.1 KiB
Go
package storage
|
|
|
|
import (
|
|
"fmt"
|
|
"math/rand"
|
|
"testing"
|
|
)
|
|
|
|
func BenchmarkMergeBlockStreamsTwoSourcesWorstCase(b *testing.B) {
|
|
benchmarkMergeBlockStreams(b, benchTwoSourcesWorstCaseMPS, benchTwoSourcesWorstCaseMPSRowsPerLoop)
|
|
}
|
|
|
|
func BenchmarkMergeBlockStreamsTwoSourcesBestCase(b *testing.B) {
|
|
benchmarkMergeBlockStreams(b, benchTwoSourcesBestCaseMPS, benchTwoSourcesBestCaseMPSRowsPerLoop)
|
|
}
|
|
|
|
func BenchmarkMergeBlockStreamsFourSourcesWorstCase(b *testing.B) {
|
|
benchmarkMergeBlockStreams(b, benchFourSourcesWorstCaseMPS, benchFourSourcesWorstCaseMPSRowsPerLoop)
|
|
}
|
|
|
|
func BenchmarkMergeBlockStreamsFourSourcesBestCase(b *testing.B) {
|
|
benchmarkMergeBlockStreams(b, benchFourSourcesBestCaseMPS, benchFourSourcesBestCaseMPSRowsPerLoop)
|
|
}
|
|
|
|
func benchmarkMergeBlockStreams(b *testing.B, mps []*inmemoryPart, rowsPerLoop int64) {
|
|
var rowsMerged, rowsDeleted uint64
|
|
|
|
b.ReportAllocs()
|
|
b.SetBytes(rowsPerLoop)
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
var bsw blockStreamWriter
|
|
var mpOut inmemoryPart
|
|
bsrs := make([]*blockStreamReader, len(mps))
|
|
for i := range mps {
|
|
var bsr blockStreamReader
|
|
bsrs[i] = &bsr
|
|
}
|
|
for pb.Next() {
|
|
for i, mp := range mps {
|
|
bsrs[i].InitFromInmemoryPart(mp)
|
|
}
|
|
mpOut.Reset()
|
|
bsw.InitFromInmemoryPart(&mpOut)
|
|
if err := mergeBlockStreams(&mpOut.ph, &bsw, bsrs, nil, nil, &rowsMerged, &rowsDeleted); err != nil {
|
|
panic(fmt.Errorf("cannot merge block streams: %w", err))
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
var benchTwoSourcesWorstCaseMPS = func() []*inmemoryPart {
|
|
var rows []rawRow
|
|
var r rawRow
|
|
r.PrecisionBits = defaultPrecisionBits
|
|
for i := 0; i < maxRowsPerBlock/2-1; i++ {
|
|
r.Value = rand.NormFloat64()
|
|
r.Timestamp = rand.Int63n(1e12)
|
|
rows = append(rows, r)
|
|
}
|
|
mp := newTestInmemoryPart(rows)
|
|
return []*inmemoryPart{mp, mp}
|
|
}()
|
|
|
|
const benchTwoSourcesWorstCaseMPSRowsPerLoop = (maxRowsPerBlock - 2)
|
|
|
|
var benchTwoSourcesBestCaseMPS = func() []*inmemoryPart {
|
|
var r rawRow
|
|
var mps []*inmemoryPart
|
|
for i := 0; i < 2; i++ {
|
|
var rows []rawRow
|
|
r.PrecisionBits = defaultPrecisionBits
|
|
r.TSID.MetricID = uint64(i)
|
|
for j := 0; j < maxRowsPerBlock; j++ {
|
|
rows = append(rows, r)
|
|
}
|
|
mp := newTestInmemoryPart(rows)
|
|
mps = append(mps, mp)
|
|
}
|
|
return mps
|
|
}()
|
|
|
|
const benchTwoSourcesBestCaseMPSRowsPerLoop = 2 * maxRowsPerBlock
|
|
|
|
var benchFourSourcesWorstCaseMPS = func() []*inmemoryPart {
|
|
var rows []rawRow
|
|
var r rawRow
|
|
r.PrecisionBits = defaultPrecisionBits
|
|
for i := 0; i < maxRowsPerBlock/2-1; i++ {
|
|
r.Value = rand.NormFloat64()
|
|
r.Timestamp = rand.Int63n(1e12)
|
|
rows = append(rows, r)
|
|
}
|
|
mp := newTestInmemoryPart(rows)
|
|
return []*inmemoryPart{mp, mp, mp, mp}
|
|
}()
|
|
|
|
const benchFourSourcesWorstCaseMPSRowsPerLoop = 2 * (maxRowsPerBlock - 2)
|
|
|
|
var benchFourSourcesBestCaseMPS = func() []*inmemoryPart {
|
|
var r rawRow
|
|
var mps []*inmemoryPart
|
|
for i := 0; i < 4; i++ {
|
|
var rows []rawRow
|
|
r.PrecisionBits = defaultPrecisionBits
|
|
r.TSID.MetricID = uint64(i)
|
|
for j := 0; j < maxRowsPerBlock; j++ {
|
|
rows = append(rows, r)
|
|
}
|
|
mp := newTestInmemoryPart(rows)
|
|
mps = append(mps, mp)
|
|
}
|
|
return mps
|
|
}()
|
|
|
|
const benchFourSourcesBestCaseMPSRowsPerLoop = 4 * maxRowsPerBlock
|