package promscrape import ( "crypto/tls" "fmt" "reflect" "regexp" "testing" "time" "github.com/VictoriaMetrics/VictoriaMetrics/lib/promauth" "github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal" "github.com/VictoriaMetrics/VictoriaMetrics/lib/promrelabel" ) func TestLoadStaticConfigs(t *testing.T) { scs, err := loadStaticConfigs("testdata/file_sd.json") if err != nil { t.Fatalf("unexpected error: %s", err) } if len(scs) == 0 { t.Fatalf("expecting non-zero static configs") } // Try loading non-existing file scs, err = loadStaticConfigs("testdata/non-exsiting-file") if err == nil { t.Fatalf("expecting non-nil error") } if scs != nil { t.Fatalf("unexpected non-nil static configs: %#v", scs) } // Try loading invalid file scs, err = loadStaticConfigs("testdata/prometheus.yml") if err == nil { t.Fatalf("expecting non-nil error") } if scs != nil { t.Fatalf("unexpected non-nil static configs: %#v", scs) } } func TestLoadConfig(t *testing.T) { cfg, _, err := loadConfig("testdata/prometheus.yml") if err != nil { t.Fatalf("unexpected error: %s", err) } if cfg == nil { t.Fatalf("expecting non-nil config") } // Try loading non-existing file cfg, _, err = loadConfig("testdata/non-existing-file") if err == nil { t.Fatalf("expecting non-nil error") } if cfg != nil { t.Fatalf("unexpected non-nil config: %#v", cfg) } // Try loading invalid file cfg, _, err = loadConfig("testdata/file_sd_1.yml") if err == nil { t.Fatalf("expecting non-nil error") } if cfg != nil { t.Fatalf("unexpected non-nil config: %#v", cfg) } } func TestGetFileSDScrapeWork(t *testing.T) { data := ` scrape_configs: - job_name: foo file_sd_configs: - files: [testdata/file_sd.json] ` var cfg Config if err := cfg.parse([]byte(data), "sss"); err != nil { t.Fatalf("cannot parase data: %s", err) } sws := cfg.getFileSDScrapeWork(nil) if !equalStaticConfigForScrapeWorks(sws, sws) { t.Fatalf("unexpected non-equal static configs;\nsws:\n%#v", sws) } // Load another static config dataNew := ` scrape_configs: - job_name: foo file_sd_configs: - files: [testdata/file_sd_1.yml] ` var cfgNew Config if err := cfgNew.parse([]byte(dataNew), "sss"); err != nil { t.Fatalf("cannot parse data: %s", err) } swsNew := cfgNew.getFileSDScrapeWork(sws) if equalStaticConfigForScrapeWorks(swsNew, sws) { t.Fatalf("unexpected equal static configs;\nswsNew:\n%#v\nsws:\n%#v", swsNew, sws) } // Try loading invalid static config data = ` scrape_configs: - job_name: foo file_sd_configs: - files: [testdata/prometheus.yml] ` if err := cfg.parse([]byte(data), "sss"); err != nil { t.Fatalf("cannot parse data: %s", err) } sws = cfg.getFileSDScrapeWork(swsNew) if len(sws) != 0 { t.Fatalf("unexpected non-empty sws:\n%#v", sws) } // Empty target in static config data = ` scrape_configs: - job_name: foo file_sd_configs: - files: [testdata/empty_target_file_sd.yml] ` if err := cfg.parse([]byte(data), "sss"); err != nil { t.Fatalf("cannot parse data: %s", err) } sws = cfg.getFileSDScrapeWork(swsNew) if len(sws) != 0 { t.Fatalf("unexpected non-empty sws:\n%#v", sws) } } func getFileSDScrapeWork(data []byte, path string) ([]ScrapeWork, error) { var cfg Config if err := cfg.parse(data, path); err != nil { return nil, fmt.Errorf("cannot parse data: %s", err) } return cfg.getFileSDScrapeWork(nil), nil } func getStaticScrapeWork(data []byte, path string) ([]ScrapeWork, error) { var cfg Config if err := cfg.parse(data, path); err != nil { return nil, fmt.Errorf("cannot parse data: %s", err) } return cfg.getStaticScrapeWork(), nil } func TestGetStaticScrapeWorkFailure(t *testing.T) { f := func(data string) { t.Helper() sws, err := getStaticScrapeWork([]byte(data), "non-existing-file") if err == nil { t.Fatalf("expecting non-nil error") } if sws != nil { t.Fatalf("expecting nil sws") } } // incorrect yaml f(`foo bar baz`) // Missing job_name f(` scrape_configs: - static_configs: - targets: ["foo"] `) // Invalid scheme f(` scrape_configs: - job_name: x scheme: asdf static_configs: - targets: ["foo"] `) // Missing username in `basic_auth` f(` scrape_configs: - job_name: x basic_auth: password: sss static_configs: - targets: ["a"] `) // Both password and password_file set in `basic_auth` f(` scrape_configs: - job_name: x basic_auth: username: foobar password: sss password_file: sdfdf static_configs: - targets: ["a"] `) // Invalid password_file set in `basic_auth` f(` scrape_configs: - job_name: x basic_auth: username: foobar password_file: /non_existing_file.pass static_configs: - targets: ["a"] `) // Both `bearer_token` and `bearer_token_file` are set f(` scrape_configs: - job_name: x bearer_token: foo bearer_token_file: bar static_configs: - targets: ["a"] `) // Both `basic_auth` and `bearer_token` are set f(` scrape_configs: - job_name: x bearer_token: foo basic_auth: username: foo password: bar static_configs: - targets: ["a"] `) // Invalid `bearer_token_file` f(` scrape_configs: - job_name: x bearer_token_file: non_existing_file.bearer static_configs: - targets: ["a"] `) // non-existing ca_file f(` scrape_configs: - job_name: aa tls_config: ca_file: non/extising/file static_configs: - targets: ["s"] `) // invalid ca_file f(` scrape_configs: - job_name: aa tls_config: ca_file: testdata/prometheus.yml static_configs: - targets: ["s"] `) // non-existing cert_file f(` scrape_configs: - job_name: aa tls_config: cert_file: non/extising/file static_configs: - targets: ["s"] `) // non-existing key_file f(` scrape_configs: - job_name: aa tls_config: key_file: non/extising/file static_configs: - targets: ["s"] `) // Invalid regex in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - regex: "(" source_labels: [foo] target_label: bar static_configs: - targets: ["s"] `) // Missing target_label for action=replace in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: replace source_labels: [foo] static_configs: - targets: ["s"] `) // Missing source_labels for action=keep in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: keep static_configs: - targets: ["s"] `) // Missing source_labels for action=drop in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: drop static_configs: - targets: ["s"] `) // Missing source_labels for action=hashmod in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: hashmod target_label: bar modulus: 123 static_configs: - targets: ["s"] `) // Missing target for action=hashmod in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: hashmod source_labels: [foo] modulus: 123 static_configs: - targets: ["s"] `) // Missing modulus for action=hashmod in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: hashmod source_labels: [foo] target_label: bar static_configs: - targets: ["s"] `) // Invalid action in relabel_configs f(` scrape_configs: - job_name: aa relabel_configs: - action: foobar static_configs: - targets: ["s"] `) } func resetScrapeWorkIDs(sws []ScrapeWork) { for i := range sws { sws[i].ID = 0 } } func TestGetFileSDScrapeWorkSuccess(t *testing.T) { f := func(data string, expectedSws []ScrapeWork) { t.Helper() sws, err := getFileSDScrapeWork([]byte(data), "non-existing-file") if err != nil { t.Fatalf("unexpected error: %s", err) } resetScrapeWorkIDs(sws) // Remove `__vm_filepath` label, since its value depends on the current working dir. for i := range sws { sw := &sws[i] for j := range sw.Labels { label := &sw.Labels[j] if label.Name == "__vm_filepath" { label.Value = "" } } } if !reflect.DeepEqual(sws, expectedSws) { t.Fatalf("unexpected scrapeWork; got\n%v\nwant\n%v", sws, expectedSws) } } f(` scrape_configs: - job_name: foo static_configs: - targets: ["xxx"] `, nil) f(` scrape_configs: - job_name: foo metrics_path: /abc/de file_sd_configs: - files: ["testdata/file_sd.json", "testdata/file_sd*.yml"] `, []ScrapeWork{ { ScrapeURL: "http://host1:80/abc/de", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "host1", }, { Name: "__metrics_path__", Value: "/abc/de", }, { Name: "__scheme__", Value: "http", }, { Name: "__vm_filepath", Value: "", }, { Name: "instance", Value: "host1:80", }, { Name: "job", Value: "foo", }, { Name: "qwe", Value: "rty", }, }, AuthConfig: &promauth.Config{}, }, { ScrapeURL: "http://host2:80/abc/de", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "host2", }, { Name: "__metrics_path__", Value: "/abc/de", }, { Name: "__scheme__", Value: "http", }, { Name: "__vm_filepath", Value: "", }, { Name: "instance", Value: "host2:80", }, { Name: "job", Value: "foo", }, { Name: "qwe", Value: "rty", }, }, AuthConfig: &promauth.Config{}, }, { ScrapeURL: "http://localhost:9090/abc/de", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "localhost:9090", }, { Name: "__metrics_path__", Value: "/abc/de", }, { Name: "__scheme__", Value: "http", }, { Name: "__vm_filepath", Value: "", }, { Name: "instance", Value: "localhost:9090", }, { Name: "job", Value: "foo", }, { Name: "yml", Value: "test", }, }, AuthConfig: &promauth.Config{}, }, }) } func TestGetStaticScrapeWorkSuccess(t *testing.T) { f := func(data string, expectedSws []ScrapeWork) { t.Helper() sws, err := getStaticScrapeWork([]byte(data), "non-exsiting-file") if err != nil { t.Fatalf("unexpected error: %s", err) } resetScrapeWorkIDs(sws) if !reflect.DeepEqual(sws, expectedSws) { t.Fatalf("unexpected scrapeWork; got\n%v\nwant\n%v", sws, expectedSws) } } f(``, nil) f(` scrape_configs: - job_name: foo static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, }, AuthConfig: &promauth.Config{}, }, }) f(` global: external_labels: datacenter: foobar jobs: xxx scrape_configs: - job_name: foo static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "datacenter", Value: "foobar", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, { Name: "jobs", Value: "xxx", }, }, AuthConfig: &promauth.Config{}, }, }) f(` global: scrape_interval: 8s scrape_timeout: 34s scrape_configs: - job_name: foo scrape_interval: 543s scrape_timeout: 12s metrics_path: /foo/bar scheme: https honor_labels: true honor_timestamps: true params: p: ["x&y", "="] xaa: bearer_token: xyz static_configs: - targets: ["foo.bar", "aaa"] labels: x: y - job_name: qwer basic_auth: username: user password: pass tls_config: server_name: foobar insecure_skip_verify: true static_configs: - targets: [1.2.3.4] `, []ScrapeWork{ { ScrapeURL: "https://foo.bar:443/foo/bar?p=x%26y&p=%3D", ScrapeInterval: 543 * time.Second, ScrapeTimeout: 12 * time.Second, HonorLabels: true, HonorTimestamps: true, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar", }, { Name: "__metrics_path__", Value: "/foo/bar", }, { Name: "__param_p", Value: "x&y", }, { Name: "__scheme__", Value: "https", }, { Name: "instance", Value: "foo.bar:443", }, { Name: "job", Value: "foo", }, { Name: "x", Value: "y", }, }, AuthConfig: &promauth.Config{ Authorization: "Bearer xyz", }, }, { ScrapeURL: "https://aaa:443/foo/bar?p=x%26y&p=%3D", ScrapeInterval: 543 * time.Second, ScrapeTimeout: 12 * time.Second, HonorLabels: true, HonorTimestamps: true, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "aaa", }, { Name: "__metrics_path__", Value: "/foo/bar", }, { Name: "__param_p", Value: "x&y", }, { Name: "__scheme__", Value: "https", }, { Name: "instance", Value: "aaa:443", }, { Name: "job", Value: "foo", }, { Name: "x", Value: "y", }, }, AuthConfig: &promauth.Config{ Authorization: "Bearer xyz", }, }, { ScrapeURL: "http://1.2.3.4:80/metrics", ScrapeInterval: 8 * time.Second, ScrapeTimeout: 34 * time.Second, HonorLabels: false, HonorTimestamps: false, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "1.2.3.4", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "1.2.3.4:80", }, { Name: "job", Value: "qwer", }, }, AuthConfig: &promauth.Config{ Authorization: "Basic dXNlcjpwYXNz", TLSServerName: "foobar", TLSInsecureSkipVerify: true, }, }, }) f(` scrape_configs: - job_name: foo relabel_configs: - source_labels: [__scheme__, __address__] separator: "://" target_label: __tmp_url - source_labels: [__tmp_url, __metrics_path__] separator: "" target_label: url - action: labeldrop regex: "job|__tmp_.+" - action: drop source_labels: [__address__] regex: "drop-.*" - action: keep source_labels: [__param_x] regex: keep_me - action: labelkeep regex: "__.*|url" - action: labelmap regex: "(url)" replacement: "prefix:${1}" - action: hashmod modulus: 123 source_labels: [__address__] target_label: hash - action: replace source_labels: [__address__] target_label: foobar replacement: "" params: x: [keep_me] static_configs: - targets: ["foo.bar:1234", "drop-this-target"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics?x=keep_me", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__param_x", Value: "keep_me", }, { Name: "__scheme__", Value: "http", }, { Name: "hash", Value: "82", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "prefix:url", Value: "http://foo.bar:1234/metrics", }, }, AuthConfig: &promauth.Config{}, }, }) f(` scrape_configs: - job_name: foo scheme: https relabel_configs: - action: replace source_labels: [non-existing-label] target_label: instance replacement: fake.addr - action: replace source_labels: [__address__] target_label: foobar regex: "missing-regex" replacement: aaabbb - action: replace source_labels: [__scheme__] target_label: job - action: replace source_labels: [__scheme__] target_label: __scheme__ replacement: mailto - target_label: __metrics_path__ replacement: /abc.de - target_label: __param_a replacement: b static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "mailto://foo.bar:1234/abc.de?a=b", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/abc.de", }, { Name: "__param_a", Value: "b", }, { Name: "__scheme__", Value: "mailto", }, { Name: "instance", Value: "fake.addr", }, { Name: "job", Value: "https", }, }, AuthConfig: &promauth.Config{}, }, }) f(` scrape_configs: - job_name: foo scheme: https relabel_configs: - action: keep source_labels: [__address__] regex: "foo\\.bar:.*" - action: hashmod source_labels: [job] modulus: 4 target_label: job - action: labeldrop regex: "non-matching-regex" - action: labelkeep regex: "job|__address__" - action: labeldrop regex: "" static_configs: - targets: ["foo.bar:1234", "xyz"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "3", }, }, AuthConfig: &promauth.Config{}, }, }) prcs, err := promrelabel.ParseRelabelConfigs(nil, []promrelabel.RelabelConfig{{ SourceLabels: []string{"foo"}, TargetLabel: "abc", }}) if err != nil { t.Fatalf("unexpected error when parsing relabel configs: %s", err) } f(` scrape_configs: - job_name: foo metric_relabel_configs: - source_labels: [foo] target_label: abc static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, }, AuthConfig: &promauth.Config{}, MetricRelabelConfigs: prcs, }, }) f(` scrape_configs: - job_name: foo basic_auth: username: xyz password_file: testdata/password.txt static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, }, AuthConfig: &promauth.Config{ Authorization: "Basic eHl6OnNlY3JldC1wYXNz", }, }, }) f(` scrape_configs: - job_name: foo bearer_token_file: testdata/password.txt static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, }, AuthConfig: &promauth.Config{ Authorization: "Bearer secret-pass", }, }, }) snakeoilCert, err := tls.LoadX509KeyPair("testdata/ssl-cert-snakeoil.pem", "testdata/ssl-cert-snakeoil.key") if err != nil { t.Fatalf("cannot load snakeoil cert: %s", err) } f(` scrape_configs: - job_name: foo tls_config: cert_file: testdata/ssl-cert-snakeoil.pem key_file: testdata/ssl-cert-snakeoil.key static_configs: - targets: ["foo.bar:1234"] `, []ScrapeWork{ { ScrapeURL: "http://foo.bar:1234/metrics", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "foo.bar:1234", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "foo.bar:1234", }, { Name: "job", Value: "foo", }, }, AuthConfig: &promauth.Config{ TLSCertificate: &snakeoilCert, }, }, }) f(` global: external_labels: job: foobar foo: xx q: qwe __address__: aaasdf __param_a: jlfd scrape_configs: - job_name: aaa params: a: [b, xy] static_configs: - targets: ["a"] labels: foo: bar __param_a: c __address__: pp job: yyy `, []ScrapeWork{ { ScrapeURL: "http://pp:80/metrics?a=c&a=xy", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "pp", }, { Name: "__metrics_path__", Value: "/metrics", }, { Name: "__param_a", Value: "c", }, { Name: "__scheme__", Value: "http", }, { Name: "foo", Value: "bar", }, { Name: "instance", Value: "pp:80", }, { Name: "job", Value: "yyy", }, { Name: "q", Value: "qwe", }, }, AuthConfig: &promauth.Config{}, }, }) f(` scrape_configs: - job_name: 'snmp' static_configs: - targets: - 192.168.1.2 # SNMP device. metrics_path: /snmp params: module: [if_mib] relabel_configs: - source_labels: [__address__] target_label: __param_target - source_labels: [__param_target] target_label: instance - target_label: __address__ replacement: 127.0.0.1:9116 # The SNMP exporter's real hostname:port. `, []ScrapeWork{ { ScrapeURL: "http://127.0.0.1:9116/snmp?module=if_mib&target=192.168.1.2", ScrapeInterval: defaultScrapeInterval, ScrapeTimeout: defaultScrapeTimeout, Labels: []prompbmarshal.Label{ { Name: "__address__", Value: "127.0.0.1:9116", }, { Name: "__metrics_path__", Value: "/snmp", }, { Name: "__param_module", Value: "if_mib", }, { Name: "__param_target", Value: "192.168.1.2", }, { Name: "__scheme__", Value: "http", }, { Name: "instance", Value: "192.168.1.2", }, { Name: "job", Value: "snmp", }, }, AuthConfig: &promauth.Config{}, }, }) } var defaultRegexForRelabelConfig = regexp.MustCompile("^(.*)$") func equalStaticConfigForScrapeWorks(a, b []ScrapeWork) bool { if len(a) != len(b) { return false } for i := range a { keyA := a[i].key() keyB := b[i].key() if keyA != keyB { return false } } return true }