Merge branch 'prometheus:main' into patch-exemplar_ui

This commit is contained in:
Vanshika 2025-01-15 18:40:15 +05:30 committed by GitHub
commit b5c5565ddf
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
15 changed files with 293 additions and 110 deletions

View file

@ -2,6 +2,8 @@
## unreleased ## unreleased
* [ENHANCEMENT] promtool: Support linting of scrape interval, through lint option `too-long-scrape-interval`. #15719
## 3.1.0 / 2025-01-02 ## 3.1.0 / 2025-01-02
* [SECURITY] upgrade golang.org/x/crypto to address reported CVE-2024-45337. #15691 * [SECURITY] upgrade golang.org/x/crypto to address reported CVE-2024-45337. #15691

View file

@ -73,14 +73,19 @@ const (
// Exit code 3 is used for "one or more lint issues detected". // Exit code 3 is used for "one or more lint issues detected".
lintErrExitCode = 3 lintErrExitCode = 3
lintOptionAll = "all" lintOptionAll = "all"
lintOptionDuplicateRules = "duplicate-rules" lintOptionDuplicateRules = "duplicate-rules"
lintOptionNone = "none" lintOptionTooLongScrapeInterval = "too-long-scrape-interval"
checkHealth = "/-/healthy" lintOptionNone = "none"
checkReadiness = "/-/ready" checkHealth = "/-/healthy"
checkReadiness = "/-/ready"
) )
var lintOptions = []string{lintOptionAll, lintOptionDuplicateRules, lintOptionNone} var (
lintRulesOptions = []string{lintOptionAll, lintOptionDuplicateRules, lintOptionNone}
// Same as lintRulesOptions, but including scrape config linting options as well.
lintConfigOptions = append(append([]string{}, lintRulesOptions...), lintOptionTooLongScrapeInterval)
)
func main() { func main() {
var ( var (
@ -97,6 +102,10 @@ func main() {
app.HelpFlag.Short('h') app.HelpFlag.Short('h')
checkCmd := app.Command("check", "Check the resources for validity.") checkCmd := app.Command("check", "Check the resources for validity.")
checkLookbackDelta := checkCmd.Flag(
"query.lookback-delta",
"The server's maximum query lookback duration.",
).Default("5m").Duration()
experimental := app.Flag("experimental", "Enable experimental commands.").Bool() experimental := app.Flag("experimental", "Enable experimental commands.").Bool()
@ -113,11 +122,12 @@ func main() {
checkConfigSyntaxOnly := checkConfigCmd.Flag("syntax-only", "Only check the config file syntax, ignoring file and content validation referenced in the config").Bool() checkConfigSyntaxOnly := checkConfigCmd.Flag("syntax-only", "Only check the config file syntax, ignoring file and content validation referenced in the config").Bool()
checkConfigLint := checkConfigCmd.Flag( checkConfigLint := checkConfigCmd.Flag(
"lint", "lint",
"Linting checks to apply to the rules specified in the config. Available options are: "+strings.Join(lintOptions, ", ")+". Use --lint=none to disable linting", "Linting checks to apply to the rules/scrape configs specified in the config. Available options are: "+strings.Join(lintConfigOptions, ", ")+". Use --lint=none to disable linting",
).Default(lintOptionDuplicateRules).String() ).Default(lintOptionDuplicateRules).String()
checkConfigLintFatal := checkConfigCmd.Flag( checkConfigLintFatal := checkConfigCmd.Flag(
"lint-fatal", "lint-fatal",
"Make lint errors exit with exit code 3.").Default("false").Bool() "Make lint errors exit with exit code 3.").Default("false").Bool()
checkConfigIgnoreUnknownFields := checkConfigCmd.Flag("ignore-unknown-fields", "Ignore unknown fields in the rule groups read by the config files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default.").Default("false").Bool()
checkWebConfigCmd := checkCmd.Command("web-config", "Check if the web config files are valid or not.") checkWebConfigCmd := checkCmd.Command("web-config", "Check if the web config files are valid or not.")
webConfigFiles := checkWebConfigCmd.Arg( webConfigFiles := checkWebConfigCmd.Arg(
@ -140,11 +150,12 @@ func main() {
).ExistingFiles() ).ExistingFiles()
checkRulesLint := checkRulesCmd.Flag( checkRulesLint := checkRulesCmd.Flag(
"lint", "lint",
"Linting checks to apply. Available options are: "+strings.Join(lintOptions, ", ")+". Use --lint=none to disable linting", "Linting checks to apply. Available options are: "+strings.Join(lintRulesOptions, ", ")+". Use --lint=none to disable linting",
).Default(lintOptionDuplicateRules).String() ).Default(lintOptionDuplicateRules).String()
checkRulesLintFatal := checkRulesCmd.Flag( checkRulesLintFatal := checkRulesCmd.Flag(
"lint-fatal", "lint-fatal",
"Make lint errors exit with exit code 3.").Default("false").Bool() "Make lint errors exit with exit code 3.").Default("false").Bool()
checkRulesIgnoreUnknownFields := checkRulesCmd.Flag("ignore-unknown-fields", "Ignore unknown fields in the rule files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default.").Default("false").Bool()
checkMetricsCmd := checkCmd.Command("metrics", checkMetricsUsage) checkMetricsCmd := checkCmd.Command("metrics", checkMetricsUsage)
checkMetricsExtended := checkCmd.Flag("extended", "Print extended information related to the cardinality of the metrics.").Bool() checkMetricsExtended := checkCmd.Flag("extended", "Print extended information related to the cardinality of the metrics.").Bool()
@ -218,6 +229,7 @@ func main() {
).Required().ExistingFiles() ).Required().ExistingFiles()
testRulesDebug := testRulesCmd.Flag("debug", "Enable unit test debugging.").Default("false").Bool() testRulesDebug := testRulesCmd.Flag("debug", "Enable unit test debugging.").Default("false").Bool()
testRulesDiff := testRulesCmd.Flag("diff", "[Experimental] Print colored differential output between expected & received output.").Default("false").Bool() testRulesDiff := testRulesCmd.Flag("diff", "[Experimental] Print colored differential output between expected & received output.").Default("false").Bool()
testRulesIgnoreUnknownFields := testRulesCmd.Flag("ignore-unknown-fields", "Ignore unknown fields in the test files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default.").Default("false").Bool()
defaultDBPath := "data/" defaultDBPath := "data/"
tsdbCmd := app.Command("tsdb", "Run tsdb commands.") tsdbCmd := app.Command("tsdb", "Run tsdb commands.")
@ -339,7 +351,7 @@ func main() {
os.Exit(CheckSD(*sdConfigFile, *sdJobName, *sdTimeout, prometheus.DefaultRegisterer)) os.Exit(CheckSD(*sdConfigFile, *sdJobName, *sdTimeout, prometheus.DefaultRegisterer))
case checkConfigCmd.FullCommand(): case checkConfigCmd.FullCommand():
os.Exit(CheckConfig(*agentMode, *checkConfigSyntaxOnly, newLintConfig(*checkConfigLint, *checkConfigLintFatal), *configFiles...)) os.Exit(CheckConfig(*agentMode, *checkConfigSyntaxOnly, newConfigLintConfig(*checkConfigLint, *checkConfigLintFatal, *checkConfigIgnoreUnknownFields, model.Duration(*checkLookbackDelta)), *configFiles...))
case checkServerHealthCmd.FullCommand(): case checkServerHealthCmd.FullCommand():
os.Exit(checkErr(CheckServerStatus(serverURL, checkHealth, httpRoundTripper))) os.Exit(checkErr(CheckServerStatus(serverURL, checkHealth, httpRoundTripper)))
@ -351,7 +363,7 @@ func main() {
os.Exit(CheckWebConfig(*webConfigFiles...)) os.Exit(CheckWebConfig(*webConfigFiles...))
case checkRulesCmd.FullCommand(): case checkRulesCmd.FullCommand():
os.Exit(CheckRules(newLintConfig(*checkRulesLint, *checkRulesLintFatal), *ruleFiles...)) os.Exit(CheckRules(newRulesLintConfig(*checkRulesLint, *checkRulesLintFatal, *checkRulesIgnoreUnknownFields), *ruleFiles...))
case checkMetricsCmd.FullCommand(): case checkMetricsCmd.FullCommand():
os.Exit(CheckMetrics(*checkMetricsExtended)) os.Exit(CheckMetrics(*checkMetricsExtended))
@ -393,6 +405,7 @@ func main() {
*testRulesRun, *testRulesRun,
*testRulesDiff, *testRulesDiff,
*testRulesDebug, *testRulesDebug,
*testRulesIgnoreUnknownFields,
*testRulesFiles...), *testRulesFiles...),
) )
@ -445,16 +458,18 @@ func checkExperimental(f bool) {
var errLint = errors.New("lint error") var errLint = errors.New("lint error")
type lintConfig struct { type rulesLintConfig struct {
all bool all bool
duplicateRules bool duplicateRules bool
fatal bool fatal bool
ignoreUnknownFields bool
} }
func newLintConfig(stringVal string, fatal bool) lintConfig { func newRulesLintConfig(stringVal string, fatal, ignoreUnknownFields bool) rulesLintConfig {
items := strings.Split(stringVal, ",") items := strings.Split(stringVal, ",")
ls := lintConfig{ ls := rulesLintConfig{
fatal: fatal, fatal: fatal,
ignoreUnknownFields: ignoreUnknownFields,
} }
for _, setting := range items { for _, setting := range items {
switch setting { switch setting {
@ -464,16 +479,57 @@ func newLintConfig(stringVal string, fatal bool) lintConfig {
ls.duplicateRules = true ls.duplicateRules = true
case lintOptionNone: case lintOptionNone:
default: default:
fmt.Printf("WARNING: unknown lint option %s\n", setting) fmt.Printf("WARNING: unknown lint option: %q\n", setting)
} }
} }
return ls return ls
} }
func (ls lintConfig) lintDuplicateRules() bool { func (ls rulesLintConfig) lintDuplicateRules() bool {
return ls.all || ls.duplicateRules return ls.all || ls.duplicateRules
} }
type configLintConfig struct {
rulesLintConfig
lookbackDelta model.Duration
}
func newConfigLintConfig(optionsStr string, fatal, ignoreUnknownFields bool, lookbackDelta model.Duration) configLintConfig {
c := configLintConfig{
rulesLintConfig: rulesLintConfig{
fatal: fatal,
},
}
lintNone := false
var rulesOptions []string
for _, option := range strings.Split(optionsStr, ",") {
switch option {
case lintOptionAll, lintOptionTooLongScrapeInterval:
c.lookbackDelta = lookbackDelta
if option == lintOptionAll {
rulesOptions = append(rulesOptions, lintOptionAll)
}
case lintOptionNone:
lintNone = true
default:
rulesOptions = append(rulesOptions, option)
}
}
if lintNone {
c.lookbackDelta = 0
rulesOptions = nil
}
if len(rulesOptions) > 0 {
c.rulesLintConfig = newRulesLintConfig(strings.Join(rulesOptions, ","), fatal, ignoreUnknownFields)
}
return c
}
// CheckServerStatus - healthy & ready. // CheckServerStatus - healthy & ready.
func CheckServerStatus(serverURL *url.URL, checkEndpoint string, roundTripper http.RoundTripper) error { func CheckServerStatus(serverURL *url.URL, checkEndpoint string, roundTripper http.RoundTripper) error {
if serverURL.Scheme == "" { if serverURL.Scheme == "" {
@ -512,12 +568,12 @@ func CheckServerStatus(serverURL *url.URL, checkEndpoint string, roundTripper ht
} }
// CheckConfig validates configuration files. // CheckConfig validates configuration files.
func CheckConfig(agentMode, checkSyntaxOnly bool, lintSettings lintConfig, files ...string) int { func CheckConfig(agentMode, checkSyntaxOnly bool, lintSettings configLintConfig, files ...string) int {
failed := false failed := false
hasErrors := false hasErrors := false
for _, f := range files { for _, f := range files {
ruleFiles, err := checkConfig(agentMode, f, checkSyntaxOnly) ruleFiles, scrapeConfigs, err := checkConfig(agentMode, f, checkSyntaxOnly)
if err != nil { if err != nil {
fmt.Fprintln(os.Stderr, " FAILED:", err) fmt.Fprintln(os.Stderr, " FAILED:", err)
hasErrors = true hasErrors = true
@ -530,12 +586,12 @@ func CheckConfig(agentMode, checkSyntaxOnly bool, lintSettings lintConfig, files
} }
fmt.Println() fmt.Println()
rulesFailed, rulesHasErrors := checkRules(ruleFiles, lintSettings) if !checkSyntaxOnly {
if rulesFailed { scrapeConfigsFailed := lintScrapeConfigs(scrapeConfigs, lintSettings)
failed = rulesFailed failed = failed || scrapeConfigsFailed
} rulesFailed, rulesHaveErrors := checkRules(ruleFiles, lintSettings.rulesLintConfig)
if rulesHasErrors { failed = failed || rulesFailed
hasErrors = rulesHasErrors hasErrors = hasErrors || rulesHaveErrors
} }
} }
if failed && hasErrors { if failed && hasErrors {
@ -574,12 +630,12 @@ func checkFileExists(fn string) error {
return err return err
} }
func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]string, error) { func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]string, []*config.ScrapeConfig, error) {
fmt.Println("Checking", filename) fmt.Println("Checking", filename)
cfg, err := config.LoadFile(filename, agentMode, promslog.NewNopLogger()) cfg, err := config.LoadFile(filename, agentMode, promslog.NewNopLogger())
if err != nil { if err != nil {
return nil, err return nil, nil, err
} }
var ruleFiles []string var ruleFiles []string
@ -587,15 +643,15 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
for _, rf := range cfg.RuleFiles { for _, rf := range cfg.RuleFiles {
rfs, err := filepath.Glob(rf) rfs, err := filepath.Glob(rf)
if err != nil { if err != nil {
return nil, err return nil, nil, err
} }
// If an explicit file was given, error if it is not accessible. // If an explicit file was given, error if it is not accessible.
if !strings.Contains(rf, "*") { if !strings.Contains(rf, "*") {
if len(rfs) == 0 { if len(rfs) == 0 {
return nil, fmt.Errorf("%q does not point to an existing file", rf) return nil, nil, fmt.Errorf("%q does not point to an existing file", rf)
} }
if err := checkFileExists(rfs[0]); err != nil { if err := checkFileExists(rfs[0]); err != nil {
return nil, fmt.Errorf("error checking rule file %q: %w", rfs[0], err) return nil, nil, fmt.Errorf("error checking rule file %q: %w", rfs[0], err)
} }
} }
ruleFiles = append(ruleFiles, rfs...) ruleFiles = append(ruleFiles, rfs...)
@ -609,26 +665,26 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
var err error var err error
scfgs, err = cfg.GetScrapeConfigs() scfgs, err = cfg.GetScrapeConfigs()
if err != nil { if err != nil {
return nil, fmt.Errorf("error loading scrape configs: %w", err) return nil, nil, fmt.Errorf("error loading scrape configs: %w", err)
} }
} }
for _, scfg := range scfgs { for _, scfg := range scfgs {
if !checkSyntaxOnly && scfg.HTTPClientConfig.Authorization != nil { if !checkSyntaxOnly && scfg.HTTPClientConfig.Authorization != nil {
if err := checkFileExists(scfg.HTTPClientConfig.Authorization.CredentialsFile); err != nil { if err := checkFileExists(scfg.HTTPClientConfig.Authorization.CredentialsFile); err != nil {
return nil, fmt.Errorf("error checking authorization credentials or bearer token file %q: %w", scfg.HTTPClientConfig.Authorization.CredentialsFile, err) return nil, nil, fmt.Errorf("error checking authorization credentials or bearer token file %q: %w", scfg.HTTPClientConfig.Authorization.CredentialsFile, err)
} }
} }
if err := checkTLSConfig(scfg.HTTPClientConfig.TLSConfig, checkSyntaxOnly); err != nil { if err := checkTLSConfig(scfg.HTTPClientConfig.TLSConfig, checkSyntaxOnly); err != nil {
return nil, err return nil, nil, err
} }
for _, c := range scfg.ServiceDiscoveryConfigs { for _, c := range scfg.ServiceDiscoveryConfigs {
switch c := c.(type) { switch c := c.(type) {
case *kubernetes.SDConfig: case *kubernetes.SDConfig:
if err := checkTLSConfig(c.HTTPClientConfig.TLSConfig, checkSyntaxOnly); err != nil { if err := checkTLSConfig(c.HTTPClientConfig.TLSConfig, checkSyntaxOnly); err != nil {
return nil, err return nil, nil, err
} }
case *file.SDConfig: case *file.SDConfig:
if checkSyntaxOnly { if checkSyntaxOnly {
@ -637,17 +693,17 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
for _, file := range c.Files { for _, file := range c.Files {
files, err := filepath.Glob(file) files, err := filepath.Glob(file)
if err != nil { if err != nil {
return nil, err return nil, nil, err
} }
if len(files) != 0 { if len(files) != 0 {
for _, f := range files { for _, f := range files {
var targetGroups []*targetgroup.Group var targetGroups []*targetgroup.Group
targetGroups, err = checkSDFile(f) targetGroups, err = checkSDFile(f)
if err != nil { if err != nil {
return nil, fmt.Errorf("checking SD file %q: %w", file, err) return nil, nil, fmt.Errorf("checking SD file %q: %w", file, err)
} }
if err := checkTargetGroupsForScrapeConfig(targetGroups, scfg); err != nil { if err := checkTargetGroupsForScrapeConfig(targetGroups, scfg); err != nil {
return nil, err return nil, nil, err
} }
} }
continue continue
@ -656,7 +712,7 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
} }
case discovery.StaticConfig: case discovery.StaticConfig:
if err := checkTargetGroupsForScrapeConfig(c, scfg); err != nil { if err := checkTargetGroupsForScrapeConfig(c, scfg); err != nil {
return nil, err return nil, nil, err
} }
} }
} }
@ -673,18 +729,18 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
for _, file := range c.Files { for _, file := range c.Files {
files, err := filepath.Glob(file) files, err := filepath.Glob(file)
if err != nil { if err != nil {
return nil, err return nil, nil, err
} }
if len(files) != 0 { if len(files) != 0 {
for _, f := range files { for _, f := range files {
var targetGroups []*targetgroup.Group var targetGroups []*targetgroup.Group
targetGroups, err = checkSDFile(f) targetGroups, err = checkSDFile(f)
if err != nil { if err != nil {
return nil, fmt.Errorf("checking SD file %q: %w", file, err) return nil, nil, fmt.Errorf("checking SD file %q: %w", file, err)
} }
if err := checkTargetGroupsForAlertmanager(targetGroups, amcfg); err != nil { if err := checkTargetGroupsForAlertmanager(targetGroups, amcfg); err != nil {
return nil, err return nil, nil, err
} }
} }
continue continue
@ -693,12 +749,12 @@ func checkConfig(agentMode bool, filename string, checkSyntaxOnly bool) ([]strin
} }
case discovery.StaticConfig: case discovery.StaticConfig:
if err := checkTargetGroupsForAlertmanager(c, amcfg); err != nil { if err := checkTargetGroupsForAlertmanager(c, amcfg); err != nil {
return nil, err return nil, nil, err
} }
} }
} }
} }
return ruleFiles, nil return ruleFiles, scfgs, nil
} }
func checkTLSConfig(tlsConfig promconfig.TLSConfig, checkSyntaxOnly bool) error { func checkTLSConfig(tlsConfig promconfig.TLSConfig, checkSyntaxOnly bool) error {
@ -760,7 +816,7 @@ func checkSDFile(filename string) ([]*targetgroup.Group, error) {
} }
// CheckRules validates rule files. // CheckRules validates rule files.
func CheckRules(ls lintConfig, files ...string) int { func CheckRules(ls rulesLintConfig, files ...string) int {
failed := false failed := false
hasErrors := false hasErrors := false
if len(files) == 0 { if len(files) == 0 {
@ -780,7 +836,7 @@ func CheckRules(ls lintConfig, files ...string) int {
} }
// checkRulesFromStdin validates rule from stdin. // checkRulesFromStdin validates rule from stdin.
func checkRulesFromStdin(ls lintConfig) (bool, bool) { func checkRulesFromStdin(ls rulesLintConfig) (bool, bool) {
failed := false failed := false
hasErrors := false hasErrors := false
fmt.Println("Checking standard input") fmt.Println("Checking standard input")
@ -789,7 +845,7 @@ func checkRulesFromStdin(ls lintConfig) (bool, bool) {
fmt.Fprintln(os.Stderr, " FAILED:", err) fmt.Fprintln(os.Stderr, " FAILED:", err)
return true, true return true, true
} }
rgs, errs := rulefmt.Parse(data) rgs, errs := rulefmt.Parse(data, ls.ignoreUnknownFields)
if errs != nil { if errs != nil {
failed = true failed = true
fmt.Fprintln(os.Stderr, " FAILED:") fmt.Fprintln(os.Stderr, " FAILED:")
@ -818,12 +874,12 @@ func checkRulesFromStdin(ls lintConfig) (bool, bool) {
} }
// checkRules validates rule files. // checkRules validates rule files.
func checkRules(files []string, ls lintConfig) (bool, bool) { func checkRules(files []string, ls rulesLintConfig) (bool, bool) {
failed := false failed := false
hasErrors := false hasErrors := false
for _, f := range files { for _, f := range files {
fmt.Println("Checking", f) fmt.Println("Checking", f)
rgs, errs := rulefmt.ParseFile(f) rgs, errs := rulefmt.ParseFile(f, ls.ignoreUnknownFields)
if errs != nil { if errs != nil {
failed = true failed = true
fmt.Fprintln(os.Stderr, " FAILED:") fmt.Fprintln(os.Stderr, " FAILED:")
@ -852,7 +908,7 @@ func checkRules(files []string, ls lintConfig) (bool, bool) {
return failed, hasErrors return failed, hasErrors
} }
func checkRuleGroups(rgs *rulefmt.RuleGroups, lintSettings lintConfig) (int, []error) { func checkRuleGroups(rgs *rulefmt.RuleGroups, lintSettings rulesLintConfig) (int, []error) {
numRules := 0 numRules := 0
for _, rg := range rgs.Groups { for _, rg := range rgs.Groups {
numRules += len(rg.Rules) numRules += len(rg.Rules)
@ -876,6 +932,16 @@ func checkRuleGroups(rgs *rulefmt.RuleGroups, lintSettings lintConfig) (int, []e
return numRules, nil return numRules, nil
} }
func lintScrapeConfigs(scrapeConfigs []*config.ScrapeConfig, lintSettings configLintConfig) bool {
for _, scfg := range scrapeConfigs {
if lintSettings.lookbackDelta > 0 && scfg.ScrapeInterval >= lintSettings.lookbackDelta {
fmt.Fprintf(os.Stderr, " FAILED: too long scrape interval found, data point will be marked as stale - job: %s, interval: %s\n", scfg.JobName, scfg.ScrapeInterval)
return true
}
}
return false
}
type compareRuleType struct { type compareRuleType struct {
metric string metric string
label labels.Labels label labels.Labels

View file

@ -185,7 +185,7 @@ func TestCheckDuplicates(t *testing.T) {
c := test c := test
t.Run(c.name, func(t *testing.T) { t.Run(c.name, func(t *testing.T) {
t.Parallel() t.Parallel()
rgs, err := rulefmt.ParseFile(c.ruleFile) rgs, err := rulefmt.ParseFile(c.ruleFile, false)
require.Empty(t, err) require.Empty(t, err)
dups := checkDuplicates(rgs.Groups) dups := checkDuplicates(rgs.Groups)
require.Equal(t, c.expectedDups, dups) require.Equal(t, c.expectedDups, dups)
@ -194,7 +194,7 @@ func TestCheckDuplicates(t *testing.T) {
} }
func BenchmarkCheckDuplicates(b *testing.B) { func BenchmarkCheckDuplicates(b *testing.B) {
rgs, err := rulefmt.ParseFile("./testdata/rules_large.yml") rgs, err := rulefmt.ParseFile("./testdata/rules_large.yml", false)
require.Empty(b, err) require.Empty(b, err)
b.ResetTimer() b.ResetTimer()
@ -234,7 +234,7 @@ func TestCheckTargetConfig(t *testing.T) {
for _, test := range cases { for _, test := range cases {
t.Run(test.name, func(t *testing.T) { t.Run(test.name, func(t *testing.T) {
t.Parallel() t.Parallel()
_, err := checkConfig(false, "testdata/"+test.file, false) _, _, err := checkConfig(false, "testdata/"+test.file, false)
if test.err != "" { if test.err != "" {
require.EqualErrorf(t, err, test.err, "Expected error %q, got %q", test.err, err.Error()) require.EqualErrorf(t, err, test.err, "Expected error %q, got %q", test.err, err.Error())
return return
@ -319,7 +319,7 @@ func TestCheckConfigSyntax(t *testing.T) {
for _, test := range cases { for _, test := range cases {
t.Run(test.name, func(t *testing.T) { t.Run(test.name, func(t *testing.T) {
t.Parallel() t.Parallel()
_, err := checkConfig(false, "testdata/"+test.file, test.syntaxOnly) _, _, err := checkConfig(false, "testdata/"+test.file, test.syntaxOnly)
expectedErrMsg := test.err expectedErrMsg := test.err
if strings.Contains(runtime.GOOS, "windows") { if strings.Contains(runtime.GOOS, "windows") {
expectedErrMsg = test.errWindows expectedErrMsg = test.errWindows
@ -355,7 +355,7 @@ func TestAuthorizationConfig(t *testing.T) {
for _, test := range cases { for _, test := range cases {
t.Run(test.name, func(t *testing.T) { t.Run(test.name, func(t *testing.T) {
t.Parallel() t.Parallel()
_, err := checkConfig(false, "testdata/"+test.file, false) _, _, err := checkConfig(false, "testdata/"+test.file, false)
if test.err != "" { if test.err != "" {
require.ErrorContains(t, err, test.err, "Expected error to contain %q, got %q", test.err, err.Error()) require.ErrorContains(t, err, test.err, "Expected error to contain %q, got %q", test.err, err.Error())
return return
@ -508,7 +508,7 @@ func TestCheckRules(t *testing.T) {
defer func(v *os.File) { os.Stdin = v }(os.Stdin) defer func(v *os.File) { os.Stdin = v }(os.Stdin)
os.Stdin = r os.Stdin = r
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, false)) exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, false, false))
require.Equal(t, successExitCode, exitCode, "") require.Equal(t, successExitCode, exitCode, "")
}) })
@ -530,7 +530,7 @@ func TestCheckRules(t *testing.T) {
defer func(v *os.File) { os.Stdin = v }(os.Stdin) defer func(v *os.File) { os.Stdin = v }(os.Stdin)
os.Stdin = r os.Stdin = r
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, false)) exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, false, false))
require.Equal(t, failureExitCode, exitCode, "") require.Equal(t, failureExitCode, exitCode, "")
}) })
@ -552,7 +552,7 @@ func TestCheckRules(t *testing.T) {
defer func(v *os.File) { os.Stdin = v }(os.Stdin) defer func(v *os.File) { os.Stdin = v }(os.Stdin)
os.Stdin = r os.Stdin = r
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, true)) exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, true, false))
require.Equal(t, lintErrExitCode, exitCode, "") require.Equal(t, lintErrExitCode, exitCode, "")
}) })
} }
@ -560,23 +560,66 @@ func TestCheckRules(t *testing.T) {
func TestCheckRulesWithRuleFiles(t *testing.T) { func TestCheckRulesWithRuleFiles(t *testing.T) {
t.Run("rules-good", func(t *testing.T) { t.Run("rules-good", func(t *testing.T) {
t.Parallel() t.Parallel()
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, false), "./testdata/rules.yml") exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, false, false), "./testdata/rules.yml")
require.Equal(t, successExitCode, exitCode, "") require.Equal(t, successExitCode, exitCode, "")
}) })
t.Run("rules-bad", func(t *testing.T) { t.Run("rules-bad", func(t *testing.T) {
t.Parallel() t.Parallel()
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, false), "./testdata/rules-bad.yml") exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, false, false), "./testdata/rules-bad.yml")
require.Equal(t, failureExitCode, exitCode, "") require.Equal(t, failureExitCode, exitCode, "")
}) })
t.Run("rules-lint-fatal", func(t *testing.T) { t.Run("rules-lint-fatal", func(t *testing.T) {
t.Parallel() t.Parallel()
exitCode := CheckRules(newLintConfig(lintOptionDuplicateRules, true), "./testdata/prometheus-rules.lint.yml") exitCode := CheckRules(newRulesLintConfig(lintOptionDuplicateRules, true, false), "./testdata/prometheus-rules.lint.yml")
require.Equal(t, lintErrExitCode, exitCode, "") require.Equal(t, lintErrExitCode, exitCode, "")
}) })
} }
func TestCheckScrapeConfigs(t *testing.T) {
for _, tc := range []struct {
name string
lookbackDelta model.Duration
expectError bool
}{
{
name: "scrape interval less than lookback delta",
lookbackDelta: model.Duration(11 * time.Minute),
expectError: false,
},
{
name: "scrape interval greater than lookback delta",
lookbackDelta: model.Duration(5 * time.Minute),
expectError: true,
},
{
name: "scrape interval same as lookback delta",
lookbackDelta: model.Duration(10 * time.Minute),
expectError: true,
},
} {
t.Run(tc.name, func(t *testing.T) {
// Non-fatal linting.
code := CheckConfig(false, false, newConfigLintConfig(lintOptionTooLongScrapeInterval, false, false, tc.lookbackDelta), "./testdata/prometheus-config.lint.too_long_scrape_interval.yml")
require.Equal(t, successExitCode, code, "Non-fatal linting should return success")
// Fatal linting.
code = CheckConfig(false, false, newConfigLintConfig(lintOptionTooLongScrapeInterval, true, false, tc.lookbackDelta), "./testdata/prometheus-config.lint.too_long_scrape_interval.yml")
if tc.expectError {
require.Equal(t, lintErrExitCode, code, "Fatal linting should return error")
} else {
require.Equal(t, successExitCode, code, "Fatal linting should return success when there are no problems")
}
// Check syntax only, no linting.
code = CheckConfig(false, true, newConfigLintConfig(lintOptionTooLongScrapeInterval, true, false, tc.lookbackDelta), "./testdata/prometheus-config.lint.too_long_scrape_interval.yml")
require.Equal(t, successExitCode, code, "Fatal linting should return success when checking syntax only")
// Lint option "none" should disable linting.
code = CheckConfig(false, false, newConfigLintConfig(lintOptionNone+","+lintOptionTooLongScrapeInterval, true, false, tc.lookbackDelta), "./testdata/prometheus-config.lint.too_long_scrape_interval.yml")
require.Equal(t, successExitCode, code, `Fatal linting should return success when lint option "none" is specified`)
})
}
}
func TestTSDBDumpCommand(t *testing.T) { func TestTSDBDumpCommand(t *testing.T) {
if testing.Short() { if testing.Short() {
t.Skip("skipping test in short mode.") t.Skip("skipping test in short mode.")

View file

@ -69,7 +69,7 @@ func newRuleImporter(logger *slog.Logger, config ruleImporterConfig, apiClient q
// loadGroups parses groups from a list of recording rule files. // loadGroups parses groups from a list of recording rule files.
func (importer *ruleImporter) loadGroups(_ context.Context, filenames []string) (errs []error) { func (importer *ruleImporter) loadGroups(_ context.Context, filenames []string) (errs []error) {
groups, errs := importer.ruleManager.LoadGroups(importer.config.evalInterval, labels.Labels{}, "", nil, filenames...) groups, errs := importer.ruleManager.LoadGroups(importer.config.evalInterval, labels.Labels{}, "", nil, false, filenames...)
if errs != nil { if errs != nil {
return errs return errs
} }

View file

@ -0,0 +1,3 @@
scrape_configs:
- job_name: too_long_scrape_interval_test
scrape_interval: 10m

View file

@ -0,0 +1,33 @@
# This is the rules file. It has an extra "ownership"
# field in the second group. promtool should ignore this field
# and not return an error with --ignore-unknown-fields.
groups:
- name: alerts
namespace: "foobar"
rules:
- alert: InstanceDown
expr: up == 0
for: 5m
labels:
severity: page
annotations:
summary: "Instance {{ $labels.instance }} down"
description: "{{ $labels.instance }} of job {{ $labels.job }} has been down for more than 5 minutes."
- alert: AlwaysFiring
expr: 1
- name: rules
ownership:
service: "test"
rules:
- record: job:test:count_over_time1m
expr: sum without(instance) (count_over_time(test[1m]))
# A recording rule that doesn't depend on input series.
- record: fixed_data
expr: 1
# Subquery with default resolution test.
- record: suquery_interval_test
expr: count_over_time(up[5m:])

View file

@ -0,0 +1,21 @@
# Minimal test case to see that --ignore-unknown-fields
# is working as expected. It should not return an error
# when any extra fields are present in the rules file.
rule_files:
- rules_extrafields.yml
evaluation_interval: 1m
tests:
- name: extra ownership field test
input_series:
- series: test
values: 1
promql_expr_test:
- expr: test
eval_time: 0
exp_samples:
- value: 1
labels: test

View file

@ -46,11 +46,11 @@ import (
// RulesUnitTest does unit testing of rules based on the unit testing files provided. // RulesUnitTest does unit testing of rules based on the unit testing files provided.
// More info about the file format can be found in the docs. // More info about the file format can be found in the docs.
func RulesUnitTest(queryOpts promqltest.LazyLoaderOpts, runStrings []string, diffFlag, debug bool, files ...string) int { func RulesUnitTest(queryOpts promqltest.LazyLoaderOpts, runStrings []string, diffFlag, debug, ignoreUnknownFields bool, files ...string) int {
return RulesUnitTestResult(io.Discard, queryOpts, runStrings, diffFlag, debug, files...) return RulesUnitTestResult(io.Discard, queryOpts, runStrings, diffFlag, debug, ignoreUnknownFields, files...)
} }
func RulesUnitTestResult(results io.Writer, queryOpts promqltest.LazyLoaderOpts, runStrings []string, diffFlag, debug bool, files ...string) int { func RulesUnitTestResult(results io.Writer, queryOpts promqltest.LazyLoaderOpts, runStrings []string, diffFlag, debug, ignoreUnknownFields bool, files ...string) int {
failed := false failed := false
junit := &junitxml.JUnitXML{} junit := &junitxml.JUnitXML{}
@ -60,7 +60,7 @@ func RulesUnitTestResult(results io.Writer, queryOpts promqltest.LazyLoaderOpts,
} }
for _, f := range files { for _, f := range files {
if errs := ruleUnitTest(f, queryOpts, run, diffFlag, debug, junit.Suite(f)); errs != nil { if errs := ruleUnitTest(f, queryOpts, run, diffFlag, debug, ignoreUnknownFields, junit.Suite(f)); errs != nil {
fmt.Fprintln(os.Stderr, " FAILED:") fmt.Fprintln(os.Stderr, " FAILED:")
for _, e := range errs { for _, e := range errs {
fmt.Fprintln(os.Stderr, e.Error()) fmt.Fprintln(os.Stderr, e.Error())
@ -82,7 +82,7 @@ func RulesUnitTestResult(results io.Writer, queryOpts promqltest.LazyLoaderOpts,
return successExitCode return successExitCode
} }
func ruleUnitTest(filename string, queryOpts promqltest.LazyLoaderOpts, run *regexp.Regexp, diffFlag, debug bool, ts *junitxml.TestSuite) []error { func ruleUnitTest(filename string, queryOpts promqltest.LazyLoaderOpts, run *regexp.Regexp, diffFlag, debug, ignoreUnknownFields bool, ts *junitxml.TestSuite) []error {
b, err := os.ReadFile(filename) b, err := os.ReadFile(filename)
if err != nil { if err != nil {
ts.Abort(err) ts.Abort(err)
@ -131,7 +131,7 @@ func ruleUnitTest(filename string, queryOpts promqltest.LazyLoaderOpts, run *reg
if t.Interval == 0 { if t.Interval == 0 {
t.Interval = unitTestInp.EvaluationInterval t.Interval = unitTestInp.EvaluationInterval
} }
ers := t.test(testname, evalInterval, groupOrderMap, queryOpts, diffFlag, debug, unitTestInp.RuleFiles...) ers := t.test(testname, evalInterval, groupOrderMap, queryOpts, diffFlag, debug, ignoreUnknownFields, unitTestInp.RuleFiles...)
if ers != nil { if ers != nil {
for _, e := range ers { for _, e := range ers {
tc.Fail(e.Error()) tc.Fail(e.Error())
@ -198,7 +198,7 @@ type testGroup struct {
} }
// test performs the unit tests. // test performs the unit tests.
func (tg *testGroup) test(testname string, evalInterval time.Duration, groupOrderMap map[string]int, queryOpts promqltest.LazyLoaderOpts, diffFlag, debug bool, ruleFiles ...string) (outErr []error) { func (tg *testGroup) test(testname string, evalInterval time.Duration, groupOrderMap map[string]int, queryOpts promqltest.LazyLoaderOpts, diffFlag, debug, ignoreUnknownFields bool, ruleFiles ...string) (outErr []error) {
if debug { if debug {
testStart := time.Now() testStart := time.Now()
fmt.Printf("DEBUG: Starting test %s\n", testname) fmt.Printf("DEBUG: Starting test %s\n", testname)
@ -228,7 +228,7 @@ func (tg *testGroup) test(testname string, evalInterval time.Duration, groupOrde
Logger: promslog.NewNopLogger(), Logger: promslog.NewNopLogger(),
} }
m := rules.NewManager(opts) m := rules.NewManager(opts)
groupsMap, ers := m.LoadGroups(time.Duration(tg.Interval), tg.ExternalLabels, tg.ExternalURL, nil, ruleFiles...) groupsMap, ers := m.LoadGroups(time.Duration(tg.Interval), tg.ExternalLabels, tg.ExternalURL, nil, ignoreUnknownFields, ruleFiles...)
if ers != nil { if ers != nil {
return ers return ers
} }

View file

@ -143,7 +143,7 @@ func TestRulesUnitTest(t *testing.T) {
} }
t.Run(tt.name, func(t *testing.T) { t.Run(tt.name, func(t *testing.T) {
t.Parallel() t.Parallel()
if got := RulesUnitTest(tt.queryOpts, nil, false, false, tt.args.files...); got != tt.want { if got := RulesUnitTest(tt.queryOpts, nil, false, false, false, tt.args.files...); got != tt.want {
t.Errorf("RulesUnitTest() = %v, want %v", got, tt.want) t.Errorf("RulesUnitTest() = %v, want %v", got, tt.want)
} }
}) })
@ -151,7 +151,7 @@ func TestRulesUnitTest(t *testing.T) {
t.Run("Junit xml output ", func(t *testing.T) { t.Run("Junit xml output ", func(t *testing.T) {
t.Parallel() t.Parallel()
var buf bytes.Buffer var buf bytes.Buffer
if got := RulesUnitTestResult(&buf, promqltest.LazyLoaderOpts{}, nil, false, false, reuseFiles...); got != 1 { if got := RulesUnitTestResult(&buf, promqltest.LazyLoaderOpts{}, nil, false, false, false, reuseFiles...); got != 1 {
t.Errorf("RulesUnitTestResults() = %v, want 1", got) t.Errorf("RulesUnitTestResults() = %v, want 1", got)
} }
var test junitxml.JUnitXML var test junitxml.JUnitXML
@ -194,10 +194,11 @@ func TestRulesUnitTestRun(t *testing.T) {
files []string files []string
} }
tests := []struct { tests := []struct {
name string name string
args args args args
queryOpts promqltest.LazyLoaderOpts queryOpts promqltest.LazyLoaderOpts
want int want int
ignoreUnknownFields bool
}{ }{
{ {
name: "Test all without run arg", name: "Test all without run arg",
@ -231,11 +232,19 @@ func TestRulesUnitTestRun(t *testing.T) {
}, },
want: 1, want: 1,
}, },
{
name: "Test all with extra fields",
args: args{
files: []string{"./testdata/rules_run_extrafields.yml"},
},
ignoreUnknownFields: true,
want: 0,
},
} }
for _, tt := range tests { for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) { t.Run(tt.name, func(t *testing.T) {
t.Parallel() t.Parallel()
got := RulesUnitTest(tt.queryOpts, tt.args.run, false, false, tt.args.files...) got := RulesUnitTest(tt.queryOpts, tt.args.run, false, false, tt.ignoreUnknownFields, tt.args.files...)
require.Equal(t, tt.want, got) require.Equal(t, tt.want, got)
}) })
} }

View file

@ -59,9 +59,10 @@ Check the resources for validity.
#### Flags #### Flags
| Flag | Description | | Flag | Description | Default |
| --- | --- | | --- | --- | --- |
| <code class="text-nowrap">--extended</code> | Print extended information related to the cardinality of the metrics. | | <code class="text-nowrap">--query.lookback-delta</code> | The server's maximum query lookback duration. | `5m` |
| <code class="text-nowrap">--extended</code> | Print extended information related to the cardinality of the metrics. | |
@ -102,8 +103,9 @@ Check if the config files are valid or not.
| Flag | Description | Default | | Flag | Description | Default |
| --- | --- | --- | | --- | --- | --- |
| <code class="text-nowrap">--syntax-only</code> | Only check the config file syntax, ignoring file and content validation referenced in the config | | | <code class="text-nowrap">--syntax-only</code> | Only check the config file syntax, ignoring file and content validation referenced in the config | |
| <code class="text-nowrap">--lint</code> | Linting checks to apply to the rules specified in the config. Available options are: all, duplicate-rules, none. Use --lint=none to disable linting | `duplicate-rules` | | <code class="text-nowrap">--lint</code> | Linting checks to apply to the rules/scrape configs specified in the config. Available options are: all, duplicate-rules, none, too-long-scrape-interval. Use --lint=none to disable linting | `duplicate-rules` |
| <code class="text-nowrap">--lint-fatal</code> | Make lint errors exit with exit code 3. | `false` | | <code class="text-nowrap">--lint-fatal</code> | Make lint errors exit with exit code 3. | `false` |
| <code class="text-nowrap">--ignore-unknown-fields</code> | Ignore unknown fields in the rule groups read by the config files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default. | `false` |
| <code class="text-nowrap">--agent</code> | Check config file for Prometheus in Agent mode. | | | <code class="text-nowrap">--agent</code> | Check config file for Prometheus in Agent mode. | |
@ -177,6 +179,7 @@ Check if the rule files are valid or not.
| --- | --- | --- | | --- | --- | --- |
| <code class="text-nowrap">--lint</code> | Linting checks to apply. Available options are: all, duplicate-rules, none. Use --lint=none to disable linting | `duplicate-rules` | | <code class="text-nowrap">--lint</code> | Linting checks to apply. Available options are: all, duplicate-rules, none. Use --lint=none to disable linting | `duplicate-rules` |
| <code class="text-nowrap">--lint-fatal</code> | Make lint errors exit with exit code 3. | `false` | | <code class="text-nowrap">--lint-fatal</code> | Make lint errors exit with exit code 3. | `false` |
| <code class="text-nowrap">--ignore-unknown-fields</code> | Ignore unknown fields in the rule files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default. | `false` |
@ -464,6 +467,7 @@ Unit tests for rules.
| <code class="text-nowrap">--run</code> <code class="text-nowrap">...<code class="text-nowrap"> | If set, will only run test groups whose names match the regular expression. Can be specified multiple times. | | | <code class="text-nowrap">--run</code> <code class="text-nowrap">...<code class="text-nowrap"> | If set, will only run test groups whose names match the regular expression. Can be specified multiple times. | |
| <code class="text-nowrap">--debug</code> | Enable unit test debugging. | `false` | | <code class="text-nowrap">--debug</code> | Enable unit test debugging. | `false` |
| <code class="text-nowrap">--diff</code> | [Experimental] Print colored differential output between expected & received output. | `false` | | <code class="text-nowrap">--diff</code> | [Experimental] Print colored differential output between expected & received output. | `false` |
| <code class="text-nowrap">--ignore-unknown-fields</code> | Ignore unknown fields in the test files. This is useful when you want to extend rule files with custom metadata. Ensure that those fields are removed before loading them into the Prometheus server as it performs strict checks by default. | `false` |

View file

@ -314,7 +314,7 @@ func testTemplateParsing(rl *RuleNode) (errs []error) {
} }
// Parse parses and validates a set of rules. // Parse parses and validates a set of rules.
func Parse(content []byte) (*RuleGroups, []error) { func Parse(content []byte, ignoreUnknownFields bool) (*RuleGroups, []error) {
var ( var (
groups RuleGroups groups RuleGroups
node ruleGroups node ruleGroups
@ -322,7 +322,9 @@ func Parse(content []byte) (*RuleGroups, []error) {
) )
decoder := yaml.NewDecoder(bytes.NewReader(content)) decoder := yaml.NewDecoder(bytes.NewReader(content))
decoder.KnownFields(true) if !ignoreUnknownFields {
decoder.KnownFields(true)
}
err := decoder.Decode(&groups) err := decoder.Decode(&groups)
// Ignore io.EOF which happens with empty input. // Ignore io.EOF which happens with empty input.
if err != nil && !errors.Is(err, io.EOF) { if err != nil && !errors.Is(err, io.EOF) {
@ -341,12 +343,12 @@ func Parse(content []byte) (*RuleGroups, []error) {
} }
// ParseFile reads and parses rules from a file. // ParseFile reads and parses rules from a file.
func ParseFile(file string) (*RuleGroups, []error) { func ParseFile(file string, ignoreUnknownFields bool) (*RuleGroups, []error) {
b, err := os.ReadFile(file) b, err := os.ReadFile(file)
if err != nil { if err != nil {
return nil, []error{fmt.Errorf("%s: %w", file, err)} return nil, []error{fmt.Errorf("%s: %w", file, err)}
} }
rgs, errs := Parse(b) rgs, errs := Parse(b, ignoreUnknownFields)
for i := range errs { for i := range errs {
errs[i] = fmt.Errorf("%s: %w", file, errs[i]) errs[i] = fmt.Errorf("%s: %w", file, errs[i])
} }

View file

@ -24,7 +24,7 @@ import (
) )
func TestParseFileSuccess(t *testing.T) { func TestParseFileSuccess(t *testing.T) {
_, errs := ParseFile("testdata/test.yaml") _, errs := ParseFile("testdata/test.yaml", false)
require.Empty(t, errs, "unexpected errors parsing file") require.Empty(t, errs, "unexpected errors parsing file")
} }
@ -84,7 +84,7 @@ func TestParseFileFailure(t *testing.T) {
} }
for _, c := range table { for _, c := range table {
_, errs := ParseFile(filepath.Join("testdata", c.filename)) _, errs := ParseFile(filepath.Join("testdata", c.filename), false)
require.NotEmpty(t, errs, "Expected error parsing %s but got none", c.filename) require.NotEmpty(t, errs, "Expected error parsing %s but got none", c.filename)
require.ErrorContainsf(t, errs[0], c.errMsg, "Expected error for %s.", c.filename) require.ErrorContainsf(t, errs[0], c.errMsg, "Expected error for %s.", c.filename)
} }
@ -179,7 +179,7 @@ groups:
} }
for _, tst := range tests { for _, tst := range tests {
rgs, errs := Parse([]byte(tst.ruleString)) rgs, errs := Parse([]byte(tst.ruleString), false)
require.NotNil(t, rgs, "Rule parsing, rule=\n"+tst.ruleString) require.NotNil(t, rgs, "Rule parsing, rule=\n"+tst.ruleString)
passed := (tst.shouldPass && len(errs) == 0) || (!tst.shouldPass && len(errs) > 0) passed := (tst.shouldPass && len(errs) == 0) || (!tst.shouldPass && len(errs) > 0)
require.True(t, passed, "Rule validation failed, rule=\n"+tst.ruleString) require.True(t, passed, "Rule validation failed, rule=\n"+tst.ruleString)
@ -206,7 +206,7 @@ groups:
annotations: annotations:
summary: "Instance {{ $labels.instance }} up" summary: "Instance {{ $labels.instance }} up"
` `
_, errs := Parse([]byte(group)) _, errs := Parse([]byte(group), false)
require.Len(t, errs, 2, "Expected two errors") require.Len(t, errs, 2, "Expected two errors")
var err00 *Error var err00 *Error
require.ErrorAs(t, errs[0], &err00) require.ErrorAs(t, errs[0], &err00)

View file

@ -207,7 +207,7 @@ func (m *Manager) Update(interval time.Duration, files []string, externalLabels
default: default:
} }
groups, errs := m.LoadGroups(interval, externalLabels, externalURL, groupEvalIterationFunc, files...) groups, errs := m.LoadGroups(interval, externalLabels, externalURL, groupEvalIterationFunc, false, files...)
if errs != nil { if errs != nil {
for _, e := range errs { for _, e := range errs {
@ -276,7 +276,7 @@ func (m *Manager) Update(interval time.Duration, files []string, externalLabels
// GroupLoader is responsible for loading rule groups from arbitrary sources and parsing them. // GroupLoader is responsible for loading rule groups from arbitrary sources and parsing them.
type GroupLoader interface { type GroupLoader interface {
Load(identifier string) (*rulefmt.RuleGroups, []error) Load(identifier string, ignoreUnknownFields bool) (*rulefmt.RuleGroups, []error)
Parse(query string) (parser.Expr, error) Parse(query string) (parser.Expr, error)
} }
@ -284,22 +284,22 @@ type GroupLoader interface {
// and parser.ParseExpr. // and parser.ParseExpr.
type FileLoader struct{} type FileLoader struct{}
func (FileLoader) Load(identifier string) (*rulefmt.RuleGroups, []error) { func (FileLoader) Load(identifier string, ignoreUnknownFields bool) (*rulefmt.RuleGroups, []error) {
return rulefmt.ParseFile(identifier) return rulefmt.ParseFile(identifier, ignoreUnknownFields)
} }
func (FileLoader) Parse(query string) (parser.Expr, error) { return parser.ParseExpr(query) } func (FileLoader) Parse(query string) (parser.Expr, error) { return parser.ParseExpr(query) }
// LoadGroups reads groups from a list of files. // LoadGroups reads groups from a list of files.
func (m *Manager) LoadGroups( func (m *Manager) LoadGroups(
interval time.Duration, externalLabels labels.Labels, externalURL string, groupEvalIterationFunc GroupEvalIterationFunc, filenames ...string, interval time.Duration, externalLabels labels.Labels, externalURL string, groupEvalIterationFunc GroupEvalIterationFunc, ignoreUnknownFields bool, filenames ...string,
) (map[string]*Group, []error) { ) (map[string]*Group, []error) {
groups := make(map[string]*Group) groups := make(map[string]*Group)
shouldRestore := !m.restored shouldRestore := !m.restored
for _, fn := range filenames { for _, fn := range filenames {
rgs, errs := m.opts.GroupLoader.Load(fn) rgs, errs := m.opts.GroupLoader.Load(fn, ignoreUnknownFields)
if errs != nil { if errs != nil {
return nil, errs return nil, errs
} }

View file

@ -808,7 +808,7 @@ func TestUpdate(t *testing.T) {
} }
// Groups will be recreated if updated. // Groups will be recreated if updated.
rgs, errs := rulefmt.ParseFile("fixtures/rules.yaml") rgs, errs := rulefmt.ParseFile("fixtures/rules.yaml", false)
require.Empty(t, errs, "file parsing failures") require.Empty(t, errs, "file parsing failures")
tmpFile, err := os.CreateTemp("", "rules.test.*.yaml") tmpFile, err := os.CreateTemp("", "rules.test.*.yaml")
@ -1532,7 +1532,7 @@ func TestManager_LoadGroups_ShouldCheckWhetherEachRuleHasDependentsAndDependenci
}) })
t.Run("load a mix of dependent and independent rules", func(t *testing.T) { t.Run("load a mix of dependent and independent rules", func(t *testing.T) {
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -1567,7 +1567,7 @@ func TestManager_LoadGroups_ShouldCheckWhetherEachRuleHasDependentsAndDependenci
}) })
t.Run("load only independent rules", func(t *testing.T) { t.Run("load only independent rules", func(t *testing.T) {
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple_independent.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple_independent.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -1975,7 +1975,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
t.Cleanup(cancel) t.Cleanup(cancel)
ruleManager := NewManager(optsFactory(storage, &maxInflight, &inflightQueries, 0)) ruleManager := NewManager(optsFactory(storage, &maxInflight, &inflightQueries, 0))
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2021,7 +2021,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2059,7 +2059,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple_independent.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple_independent.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2103,7 +2103,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple_independent.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple_independent.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2150,7 +2150,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_indeterminates.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_indeterminates.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2189,7 +2189,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_multiple_dependents_on_base.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_multiple_dependents_on_base.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
var group *Group var group *Group
@ -2235,7 +2235,7 @@ func TestAsyncRuleEvaluation(t *testing.T) {
opts.RuleConcurrencyController = nil opts.RuleConcurrencyController = nil
ruleManager := NewManager(opts) ruleManager := NewManager(opts)
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, []string{"fixtures/rules_chain.yaml"}...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, []string{"fixtures/rules_chain.yaml"}...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
var group *Group var group *Group
@ -2279,7 +2279,7 @@ func TestBoundedRuleEvalConcurrency(t *testing.T) {
ruleManager := NewManager(optsFactory(storage, &maxInflight, &inflightQueries, maxConcurrency)) ruleManager := NewManager(optsFactory(storage, &maxInflight, &inflightQueries, maxConcurrency))
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, files...) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, files...)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, groupCount) require.Len(t, groups, groupCount)
@ -2521,7 +2521,7 @@ func TestRuleDependencyController_AnalyseRules(t *testing.T) {
QueryFunc: func(ctx context.Context, q string, ts time.Time) (promql.Vector, error) { return nil, nil }, QueryFunc: func(ctx context.Context, q string, ts time.Time) (promql.Vector, error) { return nil, nil },
}) })
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, tc.ruleFile) groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, tc.ruleFile)
require.Empty(t, errs) require.Empty(t, errs)
require.Len(t, groups, 1) require.Len(t, groups, 1)
@ -2550,7 +2550,7 @@ func BenchmarkRuleDependencyController_AnalyseRules(b *testing.B) {
QueryFunc: func(ctx context.Context, q string, ts time.Time) (promql.Vector, error) { return nil, nil }, QueryFunc: func(ctx context.Context, q string, ts time.Time) (promql.Vector, error) { return nil, nil },
}) })
groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, "fixtures/rules_multiple.yaml") groups, errs := ruleManager.LoadGroups(time.Second, labels.EmptyLabels(), "", nil, false, "fixtures/rules_multiple.yaml")
require.Empty(b, errs) require.Empty(b, errs)
require.Len(b, groups, 1) require.Len(b, groups, 1)

View file

@ -1923,7 +1923,7 @@ func populateV2TimeSeries(symbolTable *writev2.SymbolsTable, batch []timeSeries,
if d.metadata != nil { if d.metadata != nil {
pendingData[nPending].Metadata.Type = writev2.FromMetadataType(d.metadata.Type) pendingData[nPending].Metadata.Type = writev2.FromMetadataType(d.metadata.Type)
pendingData[nPending].Metadata.HelpRef = symbolTable.Symbolize(d.metadata.Help) pendingData[nPending].Metadata.HelpRef = symbolTable.Symbolize(d.metadata.Help)
pendingData[nPending].Metadata.HelpRef = symbolTable.Symbolize(d.metadata.Unit) pendingData[nPending].Metadata.UnitRef = symbolTable.Symbolize(d.metadata.Unit)
nPendingMetadata++ nPendingMetadata++
} }