2019-05-22 23:16:55 +02:00
|
|
|
package storage
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"fmt"
|
|
|
|
"math/rand"
|
|
|
|
"os"
|
|
|
|
"reflect"
|
|
|
|
"regexp"
|
|
|
|
"sort"
|
|
|
|
"testing"
|
|
|
|
"testing/quick"
|
|
|
|
"time"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TestSearchQueryMarshalUnmarshal(t *testing.T) {
|
|
|
|
rnd := rand.New(rand.NewSource(0))
|
|
|
|
typ := reflect.TypeOf(&SearchQuery{})
|
|
|
|
var buf []byte
|
|
|
|
var sq2 SearchQuery
|
|
|
|
|
|
|
|
for i := 0; i < 1000; i++ {
|
|
|
|
v, ok := quick.Value(typ, rnd)
|
|
|
|
if !ok {
|
|
|
|
t.Fatalf("cannot create random SearchQuery via testing/quick.Value")
|
|
|
|
}
|
|
|
|
sq1 := v.Interface().(*SearchQuery)
|
|
|
|
if sq1 == nil {
|
|
|
|
// Skip nil sq1.
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
buf = sq1.Marshal(buf[:0])
|
|
|
|
|
|
|
|
tail, err := sq2.Unmarshal(buf)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("cannot unmarshal SearchQuery: %s", err)
|
|
|
|
}
|
|
|
|
if len(tail) > 0 {
|
|
|
|
t.Fatalf("unexpected tail left after SearchQuery unmarshaling; tail (len=%d): %q", len(tail), tail)
|
|
|
|
}
|
|
|
|
if sq1.MinTimestamp != sq2.MinTimestamp {
|
|
|
|
t.Fatalf("unexpected MinTimestamp; got %d; want %d", sq2.MinTimestamp, sq1.MinTimestamp)
|
|
|
|
}
|
|
|
|
if sq1.MaxTimestamp != sq2.MaxTimestamp {
|
|
|
|
t.Fatalf("unexpected MaxTimestamp; got %d; want %d", sq2.MaxTimestamp, sq1.MaxTimestamp)
|
|
|
|
}
|
2019-06-20 13:37:46 +02:00
|
|
|
if len(sq1.TagFilterss) != len(sq2.TagFilterss) {
|
2019-05-22 23:16:55 +02:00
|
|
|
t.Fatalf("unexpected TagFilterss len; got %d; want %d", len(sq2.TagFilterss), len(sq1.TagFilterss))
|
|
|
|
}
|
|
|
|
for ii := range sq1.TagFilterss {
|
|
|
|
tagFilters1 := sq1.TagFilterss[ii]
|
|
|
|
tagFilters2 := sq2.TagFilterss[ii]
|
|
|
|
for j := range tagFilters1 {
|
|
|
|
tf1 := &tagFilters1[j]
|
|
|
|
tf2 := &tagFilters2[j]
|
|
|
|
if string(tf1.Key) != string(tf2.Key) {
|
|
|
|
t.Fatalf("unexpected Key on iteration %d,%d; got %X; want %X", i, j, tf2.Key, tf1.Key)
|
|
|
|
}
|
|
|
|
if string(tf1.Value) != string(tf2.Value) {
|
|
|
|
t.Fatalf("unexpected Value on iteration %d,%d; got %X; want %X", i, j, tf2.Value, tf1.Value)
|
|
|
|
}
|
|
|
|
if tf1.IsNegative != tf2.IsNegative {
|
|
|
|
t.Fatalf("unexpected IsNegative on iteration %d,%d; got %v; want %v", i, j, tf2.IsNegative, tf1.IsNegative)
|
|
|
|
}
|
|
|
|
if tf1.IsRegexp != tf2.IsRegexp {
|
|
|
|
t.Fatalf("unexpected IsRegexp on iteration %d,%d; got %v; want %v", i, j, tf2.IsRegexp, tf1.IsRegexp)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestSearch(t *testing.T) {
|
2020-11-03 13:20:31 +01:00
|
|
|
path := "TestSearch"
|
2023-04-15 08:01:20 +02:00
|
|
|
st := MustOpenStorage(path, 0, 0, 0)
|
2019-05-22 23:16:55 +02:00
|
|
|
defer func() {
|
|
|
|
st.MustClose()
|
|
|
|
if err := os.RemoveAll(path); err != nil {
|
|
|
|
t.Fatalf("cannot remove storage %q: %s", path, err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Add rows to storage.
|
|
|
|
const rowsCount = 2e4
|
|
|
|
const rowsPerBlock = 1e3
|
|
|
|
const metricGroupsCount = rowsCount / 5
|
|
|
|
|
|
|
|
mrs := make([]MetricRow, rowsCount)
|
|
|
|
var mn MetricName
|
|
|
|
mn.Tags = []Tag{
|
|
|
|
{[]byte("job"), []byte("super-service")},
|
|
|
|
{[]byte("instance"), []byte("8.8.8.8:1234")},
|
|
|
|
}
|
|
|
|
startTimestamp := timestampFromTime(time.Now())
|
2019-11-08 12:16:40 +01:00
|
|
|
startTimestamp -= startTimestamp % (1e3 * 60 * 30)
|
2019-05-22 23:16:55 +02:00
|
|
|
blockRowsCount := 0
|
|
|
|
for i := 0; i < rowsCount; i++ {
|
|
|
|
mn.MetricGroup = []byte(fmt.Sprintf("metric_%d", i%metricGroupsCount))
|
|
|
|
|
|
|
|
mr := &mrs[i]
|
|
|
|
mr.MetricNameRaw = mn.marshalRaw(nil)
|
|
|
|
mr.Timestamp = startTimestamp + int64(i)
|
|
|
|
mr.Value = float64(i)
|
|
|
|
|
|
|
|
blockRowsCount++
|
|
|
|
if blockRowsCount == rowsPerBlock {
|
2024-07-17 12:07:14 +02:00
|
|
|
st.AddRows(mrs[i-blockRowsCount+1:i+1], defaultPrecisionBits)
|
2019-05-22 23:16:55 +02:00
|
|
|
blockRowsCount = 0
|
|
|
|
}
|
|
|
|
}
|
2024-07-17 12:07:14 +02:00
|
|
|
st.AddRows(mrs[rowsCount-blockRowsCount:], defaultPrecisionBits)
|
2019-05-22 23:16:55 +02:00
|
|
|
endTimestamp := mrs[len(mrs)-1].Timestamp
|
|
|
|
|
|
|
|
// Re-open the storage in order to flush all the pending cached data.
|
|
|
|
st.MustClose()
|
2023-04-15 08:01:20 +02:00
|
|
|
st = MustOpenStorage(path, 0, 0, 0)
|
2019-05-22 23:16:55 +02:00
|
|
|
|
|
|
|
// Run search.
|
|
|
|
tr := TimeRange{
|
|
|
|
MinTimestamp: startTimestamp + int64(rowsCount)/3,
|
|
|
|
MaxTimestamp: endTimestamp - int64(rowsCount)/3,
|
|
|
|
}
|
|
|
|
|
|
|
|
t.Run("serial", func(t *testing.T) {
|
2023-09-01 09:34:16 +02:00
|
|
|
if err := testSearchInternal(st, tr, mrs); err != nil {
|
2019-05-22 23:16:55 +02:00
|
|
|
t.Fatalf("unexpected error: %s", err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("concurrent", func(t *testing.T) {
|
|
|
|
ch := make(chan error, 3)
|
|
|
|
for i := 0; i < cap(ch); i++ {
|
|
|
|
go func() {
|
2023-09-01 09:34:16 +02:00
|
|
|
ch <- testSearchInternal(st, tr, mrs)
|
2019-05-22 23:16:55 +02:00
|
|
|
}()
|
|
|
|
}
|
2020-05-15 11:12:01 +02:00
|
|
|
var firstError error
|
2019-05-22 23:16:55 +02:00
|
|
|
for i := 0; i < cap(ch); i++ {
|
|
|
|
select {
|
|
|
|
case err := <-ch:
|
2020-05-15 11:12:01 +02:00
|
|
|
if err != nil && firstError == nil {
|
|
|
|
firstError = err
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
case <-time.After(10 * time.Second):
|
|
|
|
t.Fatalf("timeout")
|
|
|
|
}
|
|
|
|
}
|
2020-05-15 11:12:01 +02:00
|
|
|
if firstError != nil {
|
|
|
|
t.Fatalf("unexpected error: %s", firstError)
|
|
|
|
}
|
2019-05-22 23:16:55 +02:00
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2023-09-01 09:34:16 +02:00
|
|
|
func testSearchInternal(st *Storage, tr TimeRange, mrs []MetricRow) error {
|
2019-05-22 23:16:55 +02:00
|
|
|
var s Search
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
// Prepare TagFilters for search.
|
|
|
|
tfs := NewTagFilters()
|
|
|
|
metricGroupRe := fmt.Sprintf(`metric_\d*%d%d`, i, i)
|
|
|
|
if err := tfs.Add(nil, []byte(metricGroupRe), false, true); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot add metricGroupRe=%q: %w", metricGroupRe, err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
if err := tfs.Add([]byte("job"), []byte("nonexisting-service"), true, false); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot add tag filter %q=%q: %w", "job", "nonexsitsing-service", err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
if err := tfs.Add([]byte("instance"), []byte(".*"), false, true); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot add tag filter %q=%q: %w", "instance", ".*", err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Build extectedMrs.
|
|
|
|
var expectedMrs []MetricRow
|
|
|
|
metricGroupRegexp := regexp.MustCompile(fmt.Sprintf("^%s$", metricGroupRe))
|
|
|
|
var mn MetricName
|
|
|
|
for j := range mrs {
|
|
|
|
mr := &mrs[j]
|
|
|
|
if mr.Timestamp < tr.MinTimestamp || mr.Timestamp > tr.MaxTimestamp {
|
|
|
|
continue
|
|
|
|
}
|
2021-05-08 16:55:44 +02:00
|
|
|
if err := mn.UnmarshalRaw(mr.MetricNameRaw); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot unmarshal MetricName: %w", err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
if !metricGroupRegexp.Match(mn.MetricGroup) {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
expectedMrs = append(expectedMrs, *mr)
|
|
|
|
}
|
|
|
|
|
2020-04-27 07:13:41 +02:00
|
|
|
type metricBlock struct {
|
|
|
|
MetricName []byte
|
|
|
|
Block *Block
|
|
|
|
}
|
|
|
|
|
2019-05-22 23:16:55 +02:00
|
|
|
// Search
|
2022-06-01 01:29:19 +02:00
|
|
|
s.Init(nil, st, []*TagFilters{tfs}, tr, 1e5, noDeadline)
|
2020-04-27 07:13:41 +02:00
|
|
|
var mbs []metricBlock
|
2019-05-22 23:16:55 +02:00
|
|
|
for s.NextMetricBlock() {
|
|
|
|
var b Block
|
2022-06-28 11:55:20 +02:00
|
|
|
s.MetricBlockRef.BlockRef.MustReadBlock(&b)
|
2019-05-22 23:16:55 +02:00
|
|
|
|
2020-04-27 07:13:41 +02:00
|
|
|
var mb metricBlock
|
|
|
|
mb.MetricName = append(mb.MetricName, s.MetricBlockRef.MetricName...)
|
2019-05-22 23:16:55 +02:00
|
|
|
mb.Block = &b
|
|
|
|
mbs = append(mbs, mb)
|
|
|
|
}
|
|
|
|
if err := s.Error(); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("search error: %w", err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
s.MustClose()
|
|
|
|
|
|
|
|
// Build foundMrs.
|
|
|
|
var foundMrs []MetricRow
|
|
|
|
for _, mb := range mbs {
|
|
|
|
rb := newTestRawBlock(mb.Block, tr)
|
|
|
|
if err := mn.Unmarshal(mb.MetricName); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot unmarshal MetricName: %w", err)
|
2019-05-22 23:16:55 +02:00
|
|
|
}
|
|
|
|
metricNameRaw := mn.marshalRaw(nil)
|
|
|
|
for i, timestamp := range rb.Timestamps {
|
|
|
|
mr := MetricRow{
|
|
|
|
MetricNameRaw: metricNameRaw,
|
|
|
|
Timestamp: timestamp,
|
|
|
|
Value: rb.Values[i],
|
|
|
|
}
|
|
|
|
foundMrs = append(foundMrs, mr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compare expectedMrs to foundMrs.
|
|
|
|
sort.Slice(expectedMrs, func(i, j int) bool {
|
|
|
|
a, b := &expectedMrs[i], &expectedMrs[j]
|
|
|
|
cmp := bytes.Compare(a.MetricNameRaw, b.MetricNameRaw)
|
|
|
|
if cmp < 0 {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
if cmp > 0 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return a.Timestamp < b.Timestamp
|
|
|
|
})
|
|
|
|
sort.Slice(foundMrs, func(i, j int) bool {
|
|
|
|
a, b := &foundMrs[i], &foundMrs[j]
|
|
|
|
cmp := bytes.Compare(a.MetricNameRaw, b.MetricNameRaw)
|
|
|
|
if cmp < 0 {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
if cmp > 0 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return a.Timestamp < b.Timestamp
|
|
|
|
})
|
|
|
|
if !reflect.DeepEqual(expectedMrs, foundMrs) {
|
|
|
|
return fmt.Errorf("unexpected rows found;\ngot\n%s\nwant\n%s", mrsToString(foundMrs), mrsToString(expectedMrs))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func mrsToString(mrs []MetricRow) string {
|
|
|
|
var bb bytes.Buffer
|
|
|
|
fmt.Fprintf(&bb, "len=%d\n", len(mrs))
|
|
|
|
for i := range mrs {
|
|
|
|
mr := &mrs[i]
|
|
|
|
fmt.Fprintf(&bb, "[%q, Timestamp=%d, Value=%f]\n", mr.MetricNameRaw, mr.Timestamp, mr.Value)
|
|
|
|
}
|
|
|
|
return bb.String()
|
|
|
|
}
|