mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-12-22 16:36:27 +01:00
0bed453737
* FEATURE: [vmagent](https://docs.victoriametrics.com/vmagent.html):
allow configuring `-remoteWrite.disableOnDiskQueue` and
`-remoteWrite.dropSamplesOnOverload` cmd-line flags per each
`-remoteWrite.url`. See this [pull
request](https://github.com/VictoriaMetrics/VictoriaMetrics/pull/6065).
Thanks to @rbizos for implementaion!
* FEATURE: [vmagent](https://docs.victoriametrics.com/vmagent.html): add
labels `path` and `url` to metrics
`vmagent_remotewrite_push_failures_total` and
`vmagent_remotewrite_samples_dropped_total`. Now number of failed pushes
and dropped samples can be tracked per `-remoteWrite.url`.
---------
Signed-off-by: hagen1778 <roman@victoriametrics.com>
Co-authored-by: Raphael Bizos <r.bizos@criteo.com>
(cherry picked from commit 87fd400dfc
)
216 lines
5.2 KiB
Go
216 lines
5.2 KiB
Go
package remotewrite
|
|
|
|
import (
|
|
"fmt"
|
|
"math"
|
|
"os"
|
|
"reflect"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/promrelabel"
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/prometheus"
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/streamaggr"
|
|
"github.com/VictoriaMetrics/metrics"
|
|
)
|
|
|
|
func TestGetLabelsHash_Distribution(t *testing.T) {
|
|
f := func(bucketsCount int) {
|
|
t.Helper()
|
|
|
|
// Distribute itemsCount hashes returned by getLabelsHash() across bucketsCount buckets.
|
|
itemsCount := 1_000 * bucketsCount
|
|
m := make([]int, bucketsCount)
|
|
var labels []prompbmarshal.Label
|
|
for i := 0; i < itemsCount; i++ {
|
|
labels = append(labels[:0], prompbmarshal.Label{
|
|
Name: "__name__",
|
|
Value: fmt.Sprintf("some_name_%d", i),
|
|
})
|
|
for j := 0; j < 10; j++ {
|
|
labels = append(labels, prompbmarshal.Label{
|
|
Name: fmt.Sprintf("label_%d", j),
|
|
Value: fmt.Sprintf("value_%d_%d", i, j),
|
|
})
|
|
}
|
|
h := getLabelsHash(labels)
|
|
m[h%uint64(bucketsCount)]++
|
|
}
|
|
|
|
// Verify that the distribution is even
|
|
expectedItemsPerBucket := itemsCount / bucketsCount
|
|
for _, n := range m {
|
|
if math.Abs(1-float64(n)/float64(expectedItemsPerBucket)) > 0.04 {
|
|
t.Fatalf("unexpected items in the bucket for %d buckets; got %d; want around %d", bucketsCount, n, expectedItemsPerBucket)
|
|
}
|
|
}
|
|
}
|
|
|
|
f(2)
|
|
f(3)
|
|
f(4)
|
|
f(5)
|
|
f(10)
|
|
}
|
|
|
|
func TestRemoteWriteContext_TryPush_ImmutableTimeseries(t *testing.T) {
|
|
f := func(streamAggrConfig, relabelConfig string, dedupInterval time.Duration, keepInput, dropInput bool, input string) {
|
|
t.Helper()
|
|
perURLRelabel, err := promrelabel.ParseRelabelConfigsData([]byte(relabelConfig))
|
|
if err != nil {
|
|
t.Fatalf("cannot load relabel configs: %s", err)
|
|
}
|
|
rcs := &relabelConfigs{
|
|
perURL: []*promrelabel.ParsedConfigs{
|
|
perURLRelabel,
|
|
},
|
|
}
|
|
allRelabelConfigs.Store(rcs)
|
|
|
|
pss := make([]*pendingSeries, 1)
|
|
pss[0] = newPendingSeries(nil, true, 0, 100)
|
|
rwctx := &remoteWriteCtx{
|
|
idx: 0,
|
|
streamAggrKeepInput: keepInput,
|
|
streamAggrDropInput: dropInput,
|
|
pss: pss,
|
|
rowsPushedAfterRelabel: metrics.GetOrCreateCounter(`foo`),
|
|
rowsDroppedByRelabel: metrics.GetOrCreateCounter(`bar`),
|
|
}
|
|
if dedupInterval > 0 {
|
|
rwctx.deduplicator = streamaggr.NewDeduplicator(nil, dedupInterval, nil)
|
|
}
|
|
|
|
if len(streamAggrConfig) > 0 {
|
|
f := createFile(t, []byte(streamAggrConfig))
|
|
sas, err := streamaggr.LoadFromFile(f.Name(), nil, nil)
|
|
if err != nil {
|
|
t.Fatalf("cannot load streamaggr configs: %s", err)
|
|
}
|
|
rwctx.sas.Store(sas)
|
|
}
|
|
|
|
inputTss := mustParsePromMetrics(input)
|
|
expectedTss := make([]prompbmarshal.TimeSeries, len(inputTss))
|
|
|
|
// copy inputTss to make sure it is not mutated during TryPush call
|
|
copy(expectedTss, inputTss)
|
|
rwctx.TryPush(inputTss, false)
|
|
|
|
if !reflect.DeepEqual(expectedTss, inputTss) {
|
|
t.Fatalf("unexpected samples;\ngot\n%v\nwant\n%v", inputTss, expectedTss)
|
|
}
|
|
}
|
|
|
|
f(`
|
|
- interval: 1m
|
|
outputs: [sum_samples]
|
|
- interval: 2m
|
|
outputs: [count_series]
|
|
`, `
|
|
- action: keep
|
|
source_labels: [env]
|
|
regex: "dev"
|
|
`, 0, false, false, `
|
|
metric{env="dev"} 10
|
|
metric{env="bar"} 20
|
|
metric{env="dev"} 15
|
|
metric{env="bar"} 25
|
|
`)
|
|
f(``, ``, time.Hour, false, false, `
|
|
metric{env="dev"} 10
|
|
metric{env="foo"} 20
|
|
metric{env="dev"} 15
|
|
metric{env="foo"} 25
|
|
`)
|
|
f(``, `
|
|
- action: keep
|
|
source_labels: [env]
|
|
regex: "dev"
|
|
`, time.Hour, false, false, `
|
|
metric{env="dev"} 10
|
|
metric{env="bar"} 20
|
|
metric{env="dev"} 15
|
|
metric{env="bar"} 25
|
|
`)
|
|
f(``, `
|
|
- action: keep
|
|
source_labels: [env]
|
|
regex: "dev"
|
|
`, time.Hour, true, false, `
|
|
metric{env="test"} 10
|
|
metric{env="dev"} 20
|
|
metric{env="foo"} 15
|
|
metric{env="dev"} 25
|
|
`)
|
|
f(``, `
|
|
- action: keep
|
|
source_labels: [env]
|
|
regex: "dev"
|
|
`, time.Hour, false, true, `
|
|
metric{env="foo"} 10
|
|
metric{env="dev"} 20
|
|
metric{env="foo"} 15
|
|
metric{env="dev"} 25
|
|
`)
|
|
f(``, `
|
|
- action: keep
|
|
source_labels: [env]
|
|
regex: "dev"
|
|
`, time.Hour, true, true, `
|
|
metric{env="dev"} 10
|
|
metric{env="test"} 20
|
|
metric{env="dev"} 15
|
|
metric{env="bar"} 25
|
|
`)
|
|
}
|
|
|
|
func mustParsePromMetrics(s string) []prompbmarshal.TimeSeries {
|
|
var rows prometheus.Rows
|
|
errLogger := func(s string) {
|
|
panic(fmt.Errorf("unexpected error when parsing Prometheus metrics: %s", s))
|
|
}
|
|
rows.UnmarshalWithErrLogger(s, errLogger)
|
|
var tss []prompbmarshal.TimeSeries
|
|
samples := make([]prompbmarshal.Sample, 0, len(rows.Rows))
|
|
for _, row := range rows.Rows {
|
|
labels := make([]prompbmarshal.Label, 0, len(row.Tags)+1)
|
|
labels = append(labels, prompbmarshal.Label{
|
|
Name: "__name__",
|
|
Value: row.Metric,
|
|
})
|
|
for _, tag := range row.Tags {
|
|
labels = append(labels, prompbmarshal.Label{
|
|
Name: tag.Key,
|
|
Value: tag.Value,
|
|
})
|
|
}
|
|
samples = append(samples, prompbmarshal.Sample{
|
|
Value: row.Value,
|
|
Timestamp: row.Timestamp,
|
|
})
|
|
ts := prompbmarshal.TimeSeries{
|
|
Labels: labels,
|
|
Samples: samples[len(samples)-1:],
|
|
}
|
|
tss = append(tss, ts)
|
|
}
|
|
return tss
|
|
}
|
|
|
|
func createFile(t *testing.T, data []byte) *os.File {
|
|
t.Helper()
|
|
f, err := os.CreateTemp("", "")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := os.WriteFile(f.Name(), data, 0644); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := f.Sync(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
return f
|
|
}
|