mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-12-19 15:06:25 +01:00
161 lines
9.0 KiB
Go
161 lines
9.0 KiB
Go
|
package remotewrite
|
||
|
|
||
|
import (
|
||
|
"flag"
|
||
|
"fmt"
|
||
|
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fasttime"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/flagutil"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
|
||
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/streamaggr"
|
||
|
"github.com/VictoriaMetrics/metrics"
|
||
|
)
|
||
|
|
||
|
var (
|
||
|
// Global config
|
||
|
streamAggrGlobalConfig = flag.String("streamAggr.config", "", "Optional path to file with stream aggregation config. "+
|
||
|
"See https://docs.victoriametrics.com/stream-aggregation/ . "+
|
||
|
"See also -streamAggr.keepInput, -streamAggr.dropInput and -streamAggr.dedupInterval")
|
||
|
streamAggrGlobalKeepInput = flag.Bool("streamAggr.keepInput", false, "Whether to keep all the input samples after the aggregation "+
|
||
|
"with -streamAggr.config. By default, only aggregates samples are dropped, while the remaining samples "+
|
||
|
"are written to remote storages write. See also -streamAggr.dropInput and https://docs.victoriametrics.com/stream-aggregation/")
|
||
|
streamAggrGlobalDropInput = flag.Bool("streamAggr.dropInput", false, "Whether to drop all the input samples after the aggregation "+
|
||
|
"with -remoteWrite.streamAggr.config. By default, only aggregates samples are dropped, while the remaining samples "+
|
||
|
"are written to remote storages write. See also -streamAggr.keepInput and https://docs.victoriametrics.com/stream-aggregation/")
|
||
|
streamAggrGlobalDedupInterval = flagutil.NewDuration("streamAggr.dedupInterval", "0s", "Input samples are de-duplicated with this interval on "+
|
||
|
"aggregator before optional aggregation with -streamAggr.config . "+
|
||
|
"See also -dedup.minScrapeInterval and https://docs.victoriametrics.com/stream-aggregation/#deduplication")
|
||
|
streamAggrGlobalIgnoreOldSamples = flag.Bool("streamAggr.ignoreOldSamples", false, "Whether to ignore input samples with old timestamps outside the "+
|
||
|
"current aggregation interval for aggregator. "+
|
||
|
"See https://docs.victoriametrics.com/stream-aggregation/#ignoring-old-samples")
|
||
|
streamAggrGlobalIgnoreFirstIntervals = flag.Int("streamAggr.ignoreFirstIntervals", 0, "Number of aggregation intervals to skip after the start for "+
|
||
|
"aggregator. Increase this value if you observe incorrect aggregation results after vmagent restarts. It could be caused by receiving unordered delayed data from "+
|
||
|
"clients pushing data into the vmagent. See https://docs.victoriametrics.com/stream-aggregation/#ignore-aggregation-intervals-on-start")
|
||
|
streamAggrGlobalDropInputLabels = flagutil.NewArrayString("streamAggr.dropInputLabels", "An optional list of labels to drop from samples for aggregator "+
|
||
|
"before stream de-duplication and aggregation . See https://docs.victoriametrics.com/stream-aggregation/#dropping-unneeded-labels")
|
||
|
|
||
|
// Per URL config
|
||
|
streamAggrConfig = flagutil.NewArrayString("remoteWrite.streamAggr.config", "Optional path to file with stream aggregation config. "+
|
||
|
"See https://docs.victoriametrics.com/stream-aggregation/ . "+
|
||
|
"See also -remoteWrite.streamAggr.keepInput, -remoteWrite.streamAggr.dropInput and -remoteWrite.streamAggr.dedupInterval")
|
||
|
streamAggrDropInput = flagutil.NewArrayBool("remoteWrite.streamAggr.dropInput", "Whether to drop all the input samples after the aggregation "+
|
||
|
"with -remoteWrite.streamAggr.config. By default, only aggregates samples are dropped, while the remaining samples "+
|
||
|
"are written to the corresponding -remoteWrite.url . See also -remoteWrite.streamAggr.keepInput and https://docs.victoriametrics.com/stream-aggregation/")
|
||
|
streamAggrKeepInput = flagutil.NewArrayBool("remoteWrite.streamAggr.keepInput", "Whether to keep all the input samples after the aggregation "+
|
||
|
"with -remoteWrite.streamAggr.config. By default, only aggregates samples are dropped, while the remaining samples "+
|
||
|
"are written to the corresponding -remoteWrite.url . See also -remoteWrite.streamAggr.dropInput and https://docs.victoriametrics.com/stream-aggregation/")
|
||
|
streamAggrDedupInterval = flagutil.NewArrayDuration("remoteWrite.streamAggr.dedupInterval", 0, "Input samples are de-duplicated with this interval before optional aggregation "+
|
||
|
"with -remoteWrite.streamAggr.config . See also -dedup.minScrapeInterval and https://docs.victoriametrics.com/stream-aggregation/#deduplication")
|
||
|
streamAggrIgnoreOldSamples = flagutil.NewArrayBool("remoteWrite.streamAggr.ignoreOldSamples", "Whether to ignore input samples with old timestamps outside the current "+
|
||
|
"aggregation interval for the corresponding -remoteWrite.streamAggr.config . "+
|
||
|
"See https://docs.victoriametrics.com/stream-aggregation/#ignoring-old-samples")
|
||
|
streamAggrIgnoreFirstIntervals = flag.Int("remoteWrite.streamAggr.ignoreFirstIntervals", 0, "Number of aggregation intervals to skip after the start. Increase this value if "+
|
||
|
"you observe incorrect aggregation results after vmagent restarts. It could be caused by receiving unordered delayed data from clients pushing data into the vmagent. "+
|
||
|
"See https://docs.victoriametrics.com/stream-aggregation/#ignore-aggregation-intervals-on-start")
|
||
|
streamAggrDropInputLabels = flagutil.NewArrayString("remoteWrite.streamAggr.dropInputLabels", "An optional list of labels to drop from samples "+
|
||
|
"before stream de-duplication and aggregation . See https://docs.victoriametrics.com/stream-aggregation/#dropping-unneeded-labels")
|
||
|
)
|
||
|
|
||
|
// CheckStreamAggrConfigs checks -remoteWrite.streamAggr.config and -streamAggr.config.
|
||
|
func CheckStreamAggrConfigs() error {
|
||
|
pushNoop := func(_ []prompbmarshal.TimeSeries) {}
|
||
|
|
||
|
if _, err := newStreamAggrConfig(-1, pushNoop); err != nil {
|
||
|
return fmt.Errorf("could not load -streamAggr.config stream aggregation config: %w", err)
|
||
|
}
|
||
|
if len(*streamAggrConfig) > len(*remoteWriteURLs) {
|
||
|
return fmt.Errorf("too many -remoteWrite.streamAggr.config args: %d; it mustn't exceed the number of -remoteWrite.url args: %d",
|
||
|
len(*streamAggrConfig), len(*remoteWriteURLs))
|
||
|
}
|
||
|
for idx := range *streamAggrConfig {
|
||
|
if _, err := newStreamAggrConfig(idx, pushNoop); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// HasAnyStreamAggrConfigured checks if any streaming aggregation config provided
|
||
|
func HasAnyStreamAggrConfigured() bool {
|
||
|
return len(*streamAggrConfig) > 0 || *streamAggrGlobalConfig != ""
|
||
|
}
|
||
|
|
||
|
func reloadStreamAggrConfigs() {
|
||
|
reloadStreamAggrConfig(-1, pushToRemoteStoragesDropFailed)
|
||
|
for idx, rwctx := range rwctxs {
|
||
|
reloadStreamAggrConfig(idx, rwctx.pushInternalTrackDropped)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func reloadStreamAggrConfig(idx int, pushFunc streamaggr.PushFunc) {
|
||
|
path, opts := getStreamAggrOpts(idx)
|
||
|
logger.Infof("reloading stream aggregation configs pointed by -remoteWrite.streamAggr.config=%q", path)
|
||
|
metrics.GetOrCreateCounter(fmt.Sprintf(`vmagent_streamaggr_config_reloads_total{path=%q}`, path)).Inc()
|
||
|
|
||
|
sasNew, err := newStreamAggrConfigWithOpts(pushFunc, path, opts)
|
||
|
if err != nil {
|
||
|
metrics.GetOrCreateCounter(fmt.Sprintf(`vmagent_streamaggr_config_reloads_errors_total{path=%q}`, path)).Inc()
|
||
|
metrics.GetOrCreateCounter(fmt.Sprintf(`vmagent_streamaggr_config_reload_successful{path=%q}`, path)).Set(0)
|
||
|
logger.Errorf("cannot reload stream aggregation config at %q; continue using the previously loaded config; error: %s", path, err)
|
||
|
return
|
||
|
}
|
||
|
|
||
|
var sas *streamaggr.Aggregators
|
||
|
if idx < 0 {
|
||
|
sas = sasGlobal.Load()
|
||
|
} else {
|
||
|
sas = rwctxs[idx].sas.Load()
|
||
|
}
|
||
|
|
||
|
if !sasNew.Equal(sas) {
|
||
|
var sasOld *streamaggr.Aggregators
|
||
|
if idx < 0 {
|
||
|
sasOld = sasGlobal.Swap(sasNew)
|
||
|
} else {
|
||
|
sasOld = rwctxs[idx].sas.Swap(sasNew)
|
||
|
}
|
||
|
sasOld.MustStop()
|
||
|
logger.Infof("successfully reloaded stream aggregation configs at %q", path)
|
||
|
} else {
|
||
|
sasNew.MustStop()
|
||
|
logger.Infof("successfully reloaded stream aggregation configs at %q", path)
|
||
|
}
|
||
|
metrics.GetOrCreateCounter(fmt.Sprintf(`vmagent_streamaggr_config_reload_successful{path=%q}`, path)).Set(1)
|
||
|
metrics.GetOrCreateCounter(fmt.Sprintf(`vmagent_streamaggr_config_reload_success_timestamp_seconds{path=%q}`, path)).Set(fasttime.UnixTimestamp())
|
||
|
}
|
||
|
|
||
|
func getStreamAggrOpts(idx int) (string, *streamaggr.Options) {
|
||
|
if idx < 0 {
|
||
|
return *streamAggrGlobalConfig, &streamaggr.Options{
|
||
|
DedupInterval: streamAggrGlobalDedupInterval.Duration(),
|
||
|
DropInputLabels: *streamAggrGlobalDropInputLabels,
|
||
|
IgnoreOldSamples: *streamAggrGlobalIgnoreOldSamples,
|
||
|
IgnoreFirstIntervals: *streamAggrGlobalIgnoreFirstIntervals,
|
||
|
}
|
||
|
}
|
||
|
opts := streamaggr.Options{
|
||
|
DedupInterval: streamAggrDedupInterval.GetOptionalArg(idx),
|
||
|
DropInputLabels: *streamAggrDropInputLabels,
|
||
|
IgnoreOldSamples: streamAggrIgnoreOldSamples.GetOptionalArg(idx),
|
||
|
IgnoreFirstIntervals: *streamAggrIgnoreFirstIntervals,
|
||
|
}
|
||
|
if len(*streamAggrConfig) == 0 {
|
||
|
return "", &opts
|
||
|
}
|
||
|
return (*streamAggrConfig)[idx], &opts
|
||
|
}
|
||
|
|
||
|
func newStreamAggrConfigWithOpts(pushFunc streamaggr.PushFunc, path string, opts *streamaggr.Options) (*streamaggr.Aggregators, error) {
|
||
|
if len(path) == 0 {
|
||
|
// Skip empty stream aggregation config.
|
||
|
return nil, nil
|
||
|
}
|
||
|
return streamaggr.LoadFromFile(path, pushFunc, opts)
|
||
|
}
|
||
|
|
||
|
func newStreamAggrConfig(idx int, pushFunc streamaggr.PushFunc) (*streamaggr.Aggregators, error) {
|
||
|
path, opts := getStreamAggrOpts(idx)
|
||
|
return newStreamAggrConfigWithOpts(pushFunc, path, opts)
|
||
|
}
|