mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-27 22:49:40 -08:00
d88aea7e6f
The prefixed target provider changed a pointerized target group that was reused in the wrapped target provider, causing an ever-increasing chain of source prefixes in target groups from the Consul target provider. We now make this bug generally impossible by switching the target group channel from pointer to value type and thus ensuring that target groups are copied before being passed on to other parts of the system. I tried to not let the depointerization leak too far outside of the channel handling (both upstream and downstream) because I tried that initially and caused some nasty bugs, which I want to minimize. Fixes https://github.com/prometheus/prometheus/issues/1083
257 lines
6.7 KiB
Go
257 lines
6.7 KiB
Go
// Copyright 2015 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package discovery
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"path/filepath"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/prometheus/common/log"
|
|
"github.com/prometheus/common/model"
|
|
"gopkg.in/fsnotify.v1"
|
|
"gopkg.in/yaml.v2"
|
|
|
|
"github.com/prometheus/prometheus/config"
|
|
)
|
|
|
|
const fileSDFilepathLabel = model.MetaLabelPrefix + "filepath"
|
|
|
|
// FileDiscovery provides service discovery functionality based
|
|
// on files that contain target groups in JSON or YAML format. Refreshing
|
|
// happens using file watches and periodic refreshes.
|
|
type FileDiscovery struct {
|
|
paths []string
|
|
watcher *fsnotify.Watcher
|
|
interval time.Duration
|
|
|
|
// lastRefresh stores which files were found during the last refresh
|
|
// and how many target groups they contained.
|
|
// This is used to detect deleted target groups.
|
|
lastRefresh map[string]int
|
|
}
|
|
|
|
// NewFileDiscovery returns a new file discovery for the given paths.
|
|
func NewFileDiscovery(conf *config.FileSDConfig) *FileDiscovery {
|
|
return &FileDiscovery{
|
|
paths: conf.Names,
|
|
interval: time.Duration(conf.RefreshInterval),
|
|
}
|
|
}
|
|
|
|
// Sources implements the TargetProvider interface.
|
|
func (fd *FileDiscovery) Sources() []string {
|
|
var srcs []string
|
|
// As we allow multiple target groups per file we have no choice
|
|
// but to parse them all.
|
|
for _, p := range fd.listFiles() {
|
|
tgroups, err := readFile(p)
|
|
if err != nil {
|
|
log.Errorf("Error reading file %q: %s", p, err)
|
|
}
|
|
for _, tg := range tgroups {
|
|
srcs = append(srcs, tg.Source)
|
|
}
|
|
}
|
|
return srcs
|
|
}
|
|
|
|
// listFiles returns a list of all files that match the configured patterns.
|
|
func (fd *FileDiscovery) listFiles() []string {
|
|
var paths []string
|
|
for _, p := range fd.paths {
|
|
files, err := filepath.Glob(p)
|
|
if err != nil {
|
|
log.Errorf("Error expanding glob %q: %s", p, err)
|
|
continue
|
|
}
|
|
paths = append(paths, files...)
|
|
}
|
|
return paths
|
|
}
|
|
|
|
// watchFiles sets watches on all full paths or directories that were configured for
|
|
// this file discovery.
|
|
func (fd *FileDiscovery) watchFiles() {
|
|
if fd.watcher == nil {
|
|
panic("no watcher configured")
|
|
}
|
|
for _, p := range fd.paths {
|
|
if idx := strings.LastIndex(p, "/"); idx > -1 {
|
|
p = p[:idx]
|
|
} else {
|
|
p = "./"
|
|
}
|
|
if err := fd.watcher.Add(p); err != nil {
|
|
log.Errorf("Error adding file watch for %q: %s", p, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Run implements the TargetProvider interface.
|
|
func (fd *FileDiscovery) Run(ch chan<- config.TargetGroup, done <-chan struct{}) {
|
|
defer close(ch)
|
|
defer fd.stop()
|
|
|
|
watcher, err := fsnotify.NewWatcher()
|
|
if err != nil {
|
|
log.Errorf("Error creating file watcher: %s", err)
|
|
return
|
|
}
|
|
fd.watcher = watcher
|
|
|
|
fd.refresh(ch)
|
|
|
|
ticker := time.NewTicker(fd.interval)
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
// Stopping has priority over refreshing. Thus we wrap the actual select
|
|
// clause to always catch done signals.
|
|
select {
|
|
case <-done:
|
|
return
|
|
default:
|
|
select {
|
|
case <-done:
|
|
return
|
|
|
|
case event := <-fd.watcher.Events:
|
|
// fsnotify sometimes sends a bunch of events without name or operation.
|
|
// It's unclear what they are and why they are sent - filter them out.
|
|
if len(event.Name) == 0 {
|
|
break
|
|
}
|
|
// Everything but a chmod requires rereading.
|
|
if event.Op^fsnotify.Chmod == 0 {
|
|
break
|
|
}
|
|
// Changes to a file can spawn various sequences of events with
|
|
// different combinations of operations. For all practical purposes
|
|
// this is inaccurate.
|
|
// The most reliable solution is to reload everything if anything happens.
|
|
fd.refresh(ch)
|
|
|
|
case <-ticker.C:
|
|
// Setting a new watch after an update might fail. Make sure we don't lose
|
|
// those files forever.
|
|
fd.refresh(ch)
|
|
|
|
case err := <-fd.watcher.Errors:
|
|
if err != nil {
|
|
log.Errorf("Error on file watch: %s", err)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// stop shuts down the file watcher.
|
|
func (fd *FileDiscovery) stop() {
|
|
log.Debugf("Stopping file discovery for %s...", fd.paths)
|
|
|
|
done := make(chan struct{})
|
|
defer close(done)
|
|
|
|
// Closing the watcher will deadlock unless all events and errors are drained.
|
|
go func() {
|
|
for {
|
|
select {
|
|
case <-fd.watcher.Errors:
|
|
case <-fd.watcher.Events:
|
|
// Drain all events and errors.
|
|
case <-done:
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
if err := fd.watcher.Close(); err != nil {
|
|
log.Errorf("Error closing file watcher for %s: %s", fd.paths, err)
|
|
}
|
|
|
|
log.Debugf("File discovery for %s stopped.", fd.paths)
|
|
}
|
|
|
|
// refresh reads all files matching the discovery's patterns and sends the respective
|
|
// updated target groups through the channel.
|
|
func (fd *FileDiscovery) refresh(ch chan<- config.TargetGroup) {
|
|
ref := map[string]int{}
|
|
for _, p := range fd.listFiles() {
|
|
tgroups, err := readFile(p)
|
|
if err != nil {
|
|
log.Errorf("Error reading file %q: %s", p, err)
|
|
// Prevent deletion down below.
|
|
ref[p] = fd.lastRefresh[p]
|
|
continue
|
|
}
|
|
for _, tg := range tgroups {
|
|
ch <- *tg
|
|
}
|
|
ref[p] = len(tgroups)
|
|
}
|
|
// Send empty updates for sources that disappeared.
|
|
for f, n := range fd.lastRefresh {
|
|
m, ok := ref[f]
|
|
if !ok || n > m {
|
|
for i := m; i < n; i++ {
|
|
ch <- config.TargetGroup{Source: fileSource(f, i)}
|
|
}
|
|
}
|
|
}
|
|
fd.lastRefresh = ref
|
|
|
|
fd.watchFiles()
|
|
}
|
|
|
|
// fileSource returns a source ID for the i-th target group in the file.
|
|
func fileSource(filename string, i int) string {
|
|
return fmt.Sprintf("%s:%d", filename, i)
|
|
}
|
|
|
|
// readFile reads a JSON or YAML list of targets groups from the file, depending on its
|
|
// file extension. It returns full configuration target groups.
|
|
func readFile(filename string) ([]*config.TargetGroup, error) {
|
|
content, err := ioutil.ReadFile(filename)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var targetGroups []*config.TargetGroup
|
|
|
|
switch ext := filepath.Ext(filename); strings.ToLower(ext) {
|
|
case ".json":
|
|
if err := json.Unmarshal(content, &targetGroups); err != nil {
|
|
return nil, err
|
|
}
|
|
case ".yml", ".yaml":
|
|
if err := yaml.Unmarshal(content, &targetGroups); err != nil {
|
|
return nil, err
|
|
}
|
|
default:
|
|
panic(fmt.Errorf("retrieval.FileDiscovery.readFile: unhandled file extension %q", ext))
|
|
}
|
|
|
|
for i, tg := range targetGroups {
|
|
tg.Source = fileSource(filename, i)
|
|
if tg.Labels == nil {
|
|
tg.Labels = model.LabelSet{}
|
|
}
|
|
tg.Labels[fileSDFilepathLabel] = model.LabelValue(filename)
|
|
}
|
|
return targetGroups, nil
|
|
}
|