2021-02-01 00:10:16 +01:00
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
2022-05-03 07:03:41 +02:00
|
|
|
"context"
|
2021-02-01 00:10:16 +01:00
|
|
|
"fmt"
|
|
|
|
"log"
|
|
|
|
"os"
|
|
|
|
"os/signal"
|
|
|
|
"strings"
|
2022-04-05 16:01:32 +02:00
|
|
|
"sync/atomic"
|
2021-02-01 00:10:16 +01:00
|
|
|
"syscall"
|
|
|
|
"time"
|
|
|
|
|
2022-11-29 22:53:28 +01:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/app/vmctl/remoteread"
|
2022-09-06 09:09:34 +02:00
|
|
|
"github.com/urfave/cli/v2"
|
|
|
|
|
2021-02-01 00:10:16 +01:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/app/vmctl/influx"
|
2021-04-08 21:58:06 +02:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/app/vmctl/opentsdb"
|
2021-02-01 00:10:16 +01:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/app/vmctl/prometheus"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/app/vmctl/vm"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/buildinfo"
|
2022-04-05 16:01:32 +02:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/common"
|
2023-02-13 19:42:58 +01:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/native/stream"
|
2021-02-01 00:10:16 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
func main() {
|
2022-01-03 20:12:01 +01:00
|
|
|
var (
|
|
|
|
err error
|
|
|
|
importer *vm.Importer
|
|
|
|
)
|
|
|
|
|
2022-05-03 07:03:41 +02:00
|
|
|
ctx, cancelCtx := context.WithCancel(context.Background())
|
2021-02-01 00:10:16 +01:00
|
|
|
start := time.Now()
|
|
|
|
app := &cli.App{
|
|
|
|
Name: "vmctl",
|
2021-04-20 10:42:17 +02:00
|
|
|
Usage: "VictoriaMetrics command-line tool",
|
2021-02-01 00:10:16 +01:00
|
|
|
Version: buildinfo.Version,
|
2022-11-01 19:50:22 +01:00
|
|
|
// Disable `-version` flag to avoid conflict with lib/buildinfo flags
|
|
|
|
// see https://github.com/urfave/cli/issues/1560
|
|
|
|
HideVersion: true,
|
2021-02-01 00:10:16 +01:00
|
|
|
Commands: []*cli.Command{
|
2021-04-08 21:58:06 +02:00
|
|
|
{
|
|
|
|
Name: "opentsdb",
|
2022-11-29 22:53:28 +01:00
|
|
|
Usage: "Migrate time series from OpenTSDB",
|
2021-04-08 21:58:06 +02:00
|
|
|
Flags: mergeFlags(globalFlags, otsdbFlags, vmFlags),
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
fmt.Println("OpenTSDB import mode")
|
|
|
|
|
|
|
|
oCfg := opentsdb.Config{
|
|
|
|
Addr: c.String(otsdbAddr),
|
|
|
|
Limit: c.Int(otsdbQueryLimit),
|
|
|
|
Offset: c.Int64(otsdbOffsetDays),
|
|
|
|
HardTS: c.Int64(otsdbHardTSStart),
|
|
|
|
Retentions: c.StringSlice(otsdbRetentions),
|
|
|
|
Filters: c.StringSlice(otsdbFilters),
|
|
|
|
Normalize: c.Bool(otsdbNormalize),
|
|
|
|
MsecsTime: c.Bool(otsdbMsecsTime),
|
|
|
|
}
|
|
|
|
otsdbClient, err := opentsdb.NewClient(oCfg)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create opentsdb client: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
vmCfg := initConfigVM(c)
|
2022-05-02 09:06:34 +02:00
|
|
|
// disable progress bars since openTSDB implementation
|
|
|
|
// does not use progress bar pool
|
|
|
|
vmCfg.DisableProgressBar = true
|
2021-04-08 21:58:06 +02:00
|
|
|
importer, err := vm.NewImporter(vmCfg)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create VM importer: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
otsdbProcessor := newOtsdbProcessor(otsdbClient, importer, c.Int(otsdbConcurrency))
|
2022-01-03 20:12:01 +01:00
|
|
|
return otsdbProcessor.run(c.Bool(globalSilent), c.Bool(globalVerbose))
|
2021-04-08 21:58:06 +02:00
|
|
|
},
|
|
|
|
},
|
2021-02-01 00:10:16 +01:00
|
|
|
{
|
|
|
|
Name: "influx",
|
2022-11-29 22:53:28 +01:00
|
|
|
Usage: "Migrate time series from InfluxDB",
|
2021-02-01 00:10:16 +01:00
|
|
|
Flags: mergeFlags(globalFlags, influxFlags, vmFlags),
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
fmt.Println("InfluxDB import mode")
|
|
|
|
|
|
|
|
iCfg := influx.Config{
|
|
|
|
Addr: c.String(influxAddr),
|
|
|
|
Username: c.String(influxUser),
|
|
|
|
Password: c.String(influxPassword),
|
|
|
|
Database: c.String(influxDB),
|
|
|
|
Retention: c.String(influxRetention),
|
|
|
|
Filter: influx.Filter{
|
|
|
|
Series: c.String(influxFilterSeries),
|
|
|
|
TimeStart: c.String(influxFilterTimeStart),
|
|
|
|
TimeEnd: c.String(influxFilterTimeEnd),
|
|
|
|
},
|
|
|
|
ChunkSize: c.Int(influxChunkSize),
|
|
|
|
}
|
|
|
|
influxClient, err := influx.NewClient(iCfg)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create influx client: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
vmCfg := initConfigVM(c)
|
2022-01-03 20:12:01 +01:00
|
|
|
importer, err = vm.NewImporter(vmCfg)
|
2021-02-01 00:10:16 +01:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create VM importer: %s", err)
|
|
|
|
}
|
|
|
|
|
2022-05-03 07:03:41 +02:00
|
|
|
processor := newInfluxProcessor(
|
|
|
|
influxClient,
|
|
|
|
importer,
|
|
|
|
c.Int(influxConcurrency),
|
2022-05-06 17:06:54 +02:00
|
|
|
c.String(influxMeasurementFieldSeparator),
|
2022-05-07 21:52:42 +02:00
|
|
|
c.Bool(influxSkipDatabaseLabel),
|
|
|
|
c.Bool(influxPrometheusMode))
|
2022-01-03 20:12:01 +01:00
|
|
|
return processor.run(c.Bool(globalSilent), c.Bool(globalVerbose))
|
2021-02-01 00:10:16 +01:00
|
|
|
},
|
|
|
|
},
|
2022-11-29 22:53:28 +01:00
|
|
|
{
|
|
|
|
Name: "remote-read",
|
|
|
|
Usage: "Migrate time series via Prometheus remote-read protocol",
|
|
|
|
Flags: mergeFlags(globalFlags, remoteReadFlags, vmFlags),
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
rr, err := remoteread.NewClient(remoteread.Config{
|
2023-01-10 23:18:49 +01:00
|
|
|
Addr: c.String(remoteReadSrcAddr),
|
|
|
|
Username: c.String(remoteReadUser),
|
|
|
|
Password: c.String(remoteReadPassword),
|
|
|
|
Timeout: c.Duration(remoteReadHTTPTimeout),
|
|
|
|
UseStream: c.Bool(remoteReadUseStream),
|
|
|
|
Headers: c.String(remoteReadHeaders),
|
|
|
|
LabelName: c.String(remoteReadFilterLabel),
|
|
|
|
LabelValue: c.String(remoteReadFilterLabelValue),
|
|
|
|
InsecureSkipVerify: c.Bool(remoteReadInsecureSkipVerify),
|
2022-11-29 22:53:28 +01:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("error create remote read client: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
vmCfg := initConfigVM(c)
|
|
|
|
|
|
|
|
importer, err := vm.NewImporter(vmCfg)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create VM importer: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
rmp := remoteReadProcessor{
|
|
|
|
src: rr,
|
|
|
|
dst: importer,
|
|
|
|
filter: remoteReadFilter{
|
|
|
|
timeStart: c.Timestamp(remoteReadFilterTimeStart),
|
|
|
|
timeEnd: c.Timestamp(remoteReadFilterTimeEnd),
|
|
|
|
chunk: c.String(remoteReadStepInterval),
|
|
|
|
},
|
|
|
|
cc: c.Int(remoteReadConcurrency),
|
|
|
|
}
|
|
|
|
return rmp.run(ctx, c.Bool(globalSilent), c.Bool(globalVerbose))
|
|
|
|
},
|
|
|
|
},
|
2021-02-01 00:10:16 +01:00
|
|
|
{
|
|
|
|
Name: "prometheus",
|
2022-11-29 22:53:28 +01:00
|
|
|
Usage: "Migrate time series from Prometheus",
|
2021-02-01 00:10:16 +01:00
|
|
|
Flags: mergeFlags(globalFlags, promFlags, vmFlags),
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
fmt.Println("Prometheus import mode")
|
|
|
|
|
|
|
|
vmCfg := initConfigVM(c)
|
2022-01-03 20:12:01 +01:00
|
|
|
importer, err = vm.NewImporter(vmCfg)
|
2021-02-01 00:10:16 +01:00
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create VM importer: %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
promCfg := prometheus.Config{
|
|
|
|
Snapshot: c.String(promSnapshot),
|
|
|
|
Filter: prometheus.Filter{
|
|
|
|
TimeMin: c.String(promFilterTimeStart),
|
|
|
|
TimeMax: c.String(promFilterTimeEnd),
|
|
|
|
Label: c.String(promFilterLabel),
|
|
|
|
LabelValue: c.String(promFilterLabelValue),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
cl, err := prometheus.NewClient(promCfg)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("failed to create prometheus client: %s", err)
|
|
|
|
}
|
|
|
|
pp := prometheusProcessor{
|
|
|
|
cl: cl,
|
|
|
|
im: importer,
|
|
|
|
cc: c.Int(promConcurrency),
|
|
|
|
}
|
2022-01-03 20:12:01 +01:00
|
|
|
return pp.run(c.Bool(globalSilent), c.Bool(globalVerbose))
|
2021-02-01 00:10:16 +01:00
|
|
|
},
|
|
|
|
},
|
|
|
|
{
|
|
|
|
Name: "vm-native",
|
|
|
|
Usage: "Migrate time series between VictoriaMetrics installations via native binary format",
|
|
|
|
Flags: vmNativeFlags,
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
fmt.Println("VictoriaMetrics Native import mode")
|
|
|
|
|
|
|
|
if c.String(vmNativeFilterMatch) == "" {
|
|
|
|
return fmt.Errorf("flag %q can't be empty", vmNativeFilterMatch)
|
|
|
|
}
|
|
|
|
|
|
|
|
p := vmNativeProcessor{
|
2022-12-06 02:18:09 +01:00
|
|
|
rateLimit: c.Int64(vmRateLimit),
|
|
|
|
interCluster: c.Bool(vmInterCluster),
|
2021-02-01 00:10:16 +01:00
|
|
|
filter: filter{
|
|
|
|
match: c.String(vmNativeFilterMatch),
|
|
|
|
timeStart: c.String(vmNativeFilterTimeStart),
|
|
|
|
timeEnd: c.String(vmNativeFilterTimeEnd),
|
2022-09-06 09:09:34 +02:00
|
|
|
chunk: c.String(vmNativeStepInterval),
|
2021-02-01 00:10:16 +01:00
|
|
|
},
|
|
|
|
src: &vmNativeClient{
|
|
|
|
addr: strings.Trim(c.String(vmNativeSrcAddr), "/"),
|
|
|
|
user: c.String(vmNativeSrcUser),
|
|
|
|
password: c.String(vmNativeSrcPassword),
|
|
|
|
},
|
|
|
|
dst: &vmNativeClient{
|
|
|
|
addr: strings.Trim(c.String(vmNativeDstAddr), "/"),
|
|
|
|
user: c.String(vmNativeDstUser),
|
|
|
|
password: c.String(vmNativeDstPassword),
|
|
|
|
extraLabels: c.StringSlice(vmExtraLabel),
|
|
|
|
},
|
|
|
|
}
|
2022-05-03 07:03:41 +02:00
|
|
|
return p.run(ctx)
|
2021-02-01 00:10:16 +01:00
|
|
|
},
|
|
|
|
},
|
2022-04-05 16:01:32 +02:00
|
|
|
{
|
|
|
|
Name: "verify-block",
|
|
|
|
Usage: "Verifies exported block with VictoriaMetrics Native format",
|
|
|
|
Flags: []cli.Flag{
|
|
|
|
&cli.BoolFlag{
|
|
|
|
Name: "gunzip",
|
|
|
|
Usage: "Use GNU zip decompression for exported block",
|
|
|
|
Value: false,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
Action: func(c *cli.Context) error {
|
|
|
|
common.StartUnmarshalWorkers()
|
|
|
|
blockPath := c.Args().First()
|
|
|
|
isBlockGzipped := c.Bool("gunzip")
|
|
|
|
if len(blockPath) == 0 {
|
|
|
|
return cli.Exit("you must provide path for exported data block", 1)
|
|
|
|
}
|
|
|
|
log.Printf("verifying block at path=%q", blockPath)
|
|
|
|
f, err := os.OpenFile(blockPath, os.O_RDONLY, 0600)
|
|
|
|
if err != nil {
|
|
|
|
return cli.Exit(fmt.Errorf("cannot open exported block at path=%q err=%w", blockPath, err), 1)
|
|
|
|
}
|
|
|
|
var blocksCount uint64
|
2023-02-13 19:42:58 +01:00
|
|
|
if err := stream.Parse(f, isBlockGzipped, func(block *stream.Block) error {
|
2022-04-05 16:01:32 +02:00
|
|
|
atomic.AddUint64(&blocksCount, 1)
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return cli.Exit(fmt.Errorf("cannot parse block at path=%q, blocksCount=%d, err=%w", blockPath, blocksCount, err), 1)
|
|
|
|
}
|
|
|
|
log.Printf("successfully verified block at path=%q, blockCount=%d", blockPath, blocksCount)
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
2021-02-01 00:10:16 +01:00
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
c := make(chan os.Signal, 1)
|
|
|
|
signal.Notify(c, syscall.SIGINT, syscall.SIGTERM)
|
|
|
|
go func() {
|
|
|
|
<-c
|
|
|
|
fmt.Println("\r- Execution cancelled")
|
2022-01-03 20:12:01 +01:00
|
|
|
if importer != nil {
|
|
|
|
importer.Close()
|
|
|
|
}
|
2022-05-03 07:03:41 +02:00
|
|
|
cancelCtx()
|
2021-02-01 00:10:16 +01:00
|
|
|
}()
|
|
|
|
|
2022-01-03 20:12:01 +01:00
|
|
|
err = app.Run(os.Args)
|
2021-02-01 00:10:16 +01:00
|
|
|
if err != nil {
|
2022-04-13 11:52:55 +02:00
|
|
|
log.Fatalln(err)
|
2021-02-01 00:10:16 +01:00
|
|
|
}
|
|
|
|
log.Printf("Total time: %v", time.Since(start))
|
|
|
|
}
|
|
|
|
|
|
|
|
func initConfigVM(c *cli.Context) vm.Config {
|
|
|
|
return vm.Config{
|
|
|
|
Addr: c.String(vmAddr),
|
|
|
|
User: c.String(vmUser),
|
|
|
|
Password: c.String(vmPassword),
|
|
|
|
Concurrency: uint8(c.Int(vmConcurrency)),
|
|
|
|
Compress: c.Bool(vmCompress),
|
|
|
|
AccountID: c.String(vmAccountID),
|
|
|
|
BatchSize: c.Int(vmBatchSize),
|
|
|
|
SignificantFigures: c.Int(vmSignificantFigures),
|
2021-02-01 13:27:05 +01:00
|
|
|
RoundDigits: c.Int(vmRoundDigits),
|
2021-02-01 00:10:16 +01:00
|
|
|
ExtraLabels: c.StringSlice(vmExtraLabel),
|
2021-12-24 11:18:07 +01:00
|
|
|
RateLimit: c.Int64(vmRateLimit),
|
2022-05-02 09:06:34 +02:00
|
|
|
DisableProgressBar: c.Bool(vmDisableProgressBar),
|
2021-02-01 00:10:16 +01:00
|
|
|
}
|
|
|
|
}
|