mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-12-15 08:23:34 +01:00
adc18c3ee6
This should reduce the amounts of used RAM during queries with filters over big number of time series.
115 lines
3.1 KiB
Go
115 lines
3.1 KiB
Go
package mergeset
|
|
|
|
import (
|
|
"bytes"
|
|
"fmt"
|
|
"math/rand"
|
|
"os"
|
|
"testing"
|
|
)
|
|
|
|
func BenchmarkTableSearch(b *testing.B) {
|
|
for _, itemsCount := range []int{1e5, 1e6, 1e7} {
|
|
b.Run(fmt.Sprintf("itemsCount-%d", itemsCount), func(b *testing.B) {
|
|
benchmarkTableSearch(b, itemsCount)
|
|
})
|
|
}
|
|
}
|
|
|
|
func benchmarkTableSearch(b *testing.B, itemsCount int) {
|
|
path := fmt.Sprintf("BenchmarkTableSearch-%d", itemsCount)
|
|
if err := os.RemoveAll(path); err != nil {
|
|
b.Fatalf("cannot remove %q: %s", path, err)
|
|
}
|
|
defer func() {
|
|
_ = os.RemoveAll(path)
|
|
}()
|
|
|
|
tb, items, err := newTestTable(path, itemsCount)
|
|
if err != nil {
|
|
panic(fmt.Errorf("cannot create test table at %q with %d items: %s", path, itemsCount, err))
|
|
}
|
|
|
|
// Force finishing pending merges
|
|
tb.MustClose()
|
|
tb, err = OpenTable(path, nil, nil)
|
|
if err != nil {
|
|
b.Fatalf("unexpected error when re-opening table %q: %s", path, err)
|
|
}
|
|
defer tb.MustClose()
|
|
|
|
keys := make([][]byte, len(items))
|
|
for i, item := range items {
|
|
keys[i] = []byte(item)
|
|
}
|
|
|
|
b.Run("sequential-keys-exact", func(b *testing.B) {
|
|
benchmarkTableSearchKeys(b, tb, keys, 0)
|
|
})
|
|
b.Run("sequential-keys-without-siffux", func(b *testing.B) {
|
|
benchmarkTableSearchKeys(b, tb, keys, 4)
|
|
})
|
|
|
|
randKeys := append([][]byte{}, keys...)
|
|
rand.Shuffle(len(randKeys), func(i, j int) {
|
|
randKeys[i], randKeys[j] = randKeys[j], randKeys[i]
|
|
})
|
|
b.Run("random-keys-exact", func(b *testing.B) {
|
|
benchmarkTableSearchKeys(b, tb, randKeys, 0)
|
|
})
|
|
b.Run("random-keys-without-siffux", func(b *testing.B) {
|
|
benchmarkTableSearchKeys(b, tb, randKeys, 4)
|
|
})
|
|
}
|
|
|
|
func benchmarkTableSearchKeys(b *testing.B, tb *Table, keys [][]byte, stripSuffix int) {
|
|
for _, rowsToScan := range []int{1, 10, 100, 1000} {
|
|
b.Run(fmt.Sprintf("rowsToScan-%d", rowsToScan), func(b *testing.B) {
|
|
benchmarkTableSearchKeysExt(b, tb, keys, stripSuffix, rowsToScan)
|
|
})
|
|
}
|
|
}
|
|
|
|
func benchmarkTableSearchKeysExt(b *testing.B, tb *Table, keys [][]byte, stripSuffix, rowsToScan int) {
|
|
searchKeysCount := 1000
|
|
if searchKeysCount >= len(keys) {
|
|
searchKeysCount = len(keys) - 1
|
|
}
|
|
|
|
b.ResetTimer()
|
|
b.ReportAllocs()
|
|
b.SetBytes(int64(searchKeysCount * rowsToScan))
|
|
b.RunParallel(func(pb *testing.PB) {
|
|
var ts TableSearch
|
|
ts.Init(tb, nil)
|
|
defer ts.MustClose()
|
|
for pb.Next() {
|
|
startIdx := rand.Intn(len(keys) - searchKeysCount)
|
|
searchKeys := keys[startIdx : startIdx+searchKeysCount]
|
|
for i, key := range searchKeys {
|
|
searchKey := key
|
|
if len(searchKey) < stripSuffix {
|
|
searchKey = nil
|
|
} else {
|
|
searchKey = searchKey[:len(searchKey)-stripSuffix]
|
|
}
|
|
ts.Seek(searchKey)
|
|
if !ts.NextItem() {
|
|
panic(fmt.Errorf("BUG: NextItem must return true for searchKeys[%d]=%q; err=%v", i, searchKey, ts.Error()))
|
|
}
|
|
if !bytes.HasPrefix(ts.Item, searchKey) {
|
|
panic(fmt.Errorf("BUG: unexpected item found for searchKey[%d]=%q; got %q; want %q", i, searchKey, ts.Item, key))
|
|
}
|
|
for j := 1; j < rowsToScan; j++ {
|
|
if !ts.NextItem() {
|
|
break
|
|
}
|
|
}
|
|
if err := ts.Error(); err != nil {
|
|
panic(fmt.Errorf("BUG: unexpected error for searchKeys[%d]=%q: %s", i, searchKey, err))
|
|
}
|
|
}
|
|
}
|
|
})
|
|
}
|