mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-30 07:59:40 -08:00
a3e9628e0c
When using Kubernetes on cloud providers, nodes will have the spec.providerID field populated to contain the cloud provider specific name of the EC2/GCE/... instance. Let's expose this information as an additional label, so that it's easier to annotate metrics and alerts to contain the cloud provider specific name of the instance to which it pertains. Signed-off-by: Ed Schouten <eschouten@apple.com>
243 lines
6.3 KiB
Go
243 lines
6.3 KiB
Go
// Copyright 2016 The Prometheus Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package kubernetes
|
|
|
|
import (
|
|
"context"
|
|
"net"
|
|
"strconv"
|
|
|
|
"github.com/go-kit/log"
|
|
"github.com/go-kit/log/level"
|
|
"github.com/pkg/errors"
|
|
"github.com/prometheus/common/model"
|
|
apiv1 "k8s.io/api/core/v1"
|
|
"k8s.io/client-go/tools/cache"
|
|
"k8s.io/client-go/util/workqueue"
|
|
|
|
"github.com/prometheus/prometheus/discovery/targetgroup"
|
|
"github.com/prometheus/prometheus/util/strutil"
|
|
)
|
|
|
|
const (
|
|
NodeLegacyHostIP = "LegacyHostIP"
|
|
)
|
|
|
|
var (
|
|
nodeAddCount = eventCount.WithLabelValues("node", "add")
|
|
nodeUpdateCount = eventCount.WithLabelValues("node", "update")
|
|
nodeDeleteCount = eventCount.WithLabelValues("node", "delete")
|
|
)
|
|
|
|
// Node discovers Kubernetes nodes.
|
|
type Node struct {
|
|
logger log.Logger
|
|
informer cache.SharedInformer
|
|
store cache.Store
|
|
queue *workqueue.Type
|
|
}
|
|
|
|
// NewNode returns a new node discovery.
|
|
func NewNode(l log.Logger, inf cache.SharedInformer) *Node {
|
|
if l == nil {
|
|
l = log.NewNopLogger()
|
|
}
|
|
n := &Node{logger: l, informer: inf, store: inf.GetStore(), queue: workqueue.NewNamed("node")}
|
|
n.informer.AddEventHandler(cache.ResourceEventHandlerFuncs{
|
|
AddFunc: func(o interface{}) {
|
|
nodeAddCount.Inc()
|
|
n.enqueue(o)
|
|
},
|
|
DeleteFunc: func(o interface{}) {
|
|
nodeDeleteCount.Inc()
|
|
n.enqueue(o)
|
|
},
|
|
UpdateFunc: func(_, o interface{}) {
|
|
nodeUpdateCount.Inc()
|
|
n.enqueue(o)
|
|
},
|
|
})
|
|
return n
|
|
}
|
|
|
|
func (n *Node) enqueue(obj interface{}) {
|
|
key, err := cache.DeletionHandlingMetaNamespaceKeyFunc(obj)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
n.queue.Add(key)
|
|
}
|
|
|
|
// Run implements the Discoverer interface.
|
|
func (n *Node) Run(ctx context.Context, ch chan<- []*targetgroup.Group) {
|
|
defer n.queue.ShutDown()
|
|
|
|
if !cache.WaitForCacheSync(ctx.Done(), n.informer.HasSynced) {
|
|
if ctx.Err() != context.Canceled {
|
|
level.Error(n.logger).Log("msg", "node informer unable to sync cache")
|
|
}
|
|
return
|
|
}
|
|
|
|
go func() {
|
|
for n.process(ctx, ch) {
|
|
}
|
|
}()
|
|
|
|
// Block until the target provider is explicitly canceled.
|
|
<-ctx.Done()
|
|
}
|
|
|
|
func (n *Node) process(ctx context.Context, ch chan<- []*targetgroup.Group) bool {
|
|
keyObj, quit := n.queue.Get()
|
|
if quit {
|
|
return false
|
|
}
|
|
defer n.queue.Done(keyObj)
|
|
key := keyObj.(string)
|
|
|
|
_, name, err := cache.SplitMetaNamespaceKey(key)
|
|
if err != nil {
|
|
return true
|
|
}
|
|
|
|
o, exists, err := n.store.GetByKey(key)
|
|
if err != nil {
|
|
return true
|
|
}
|
|
if !exists {
|
|
send(ctx, ch, &targetgroup.Group{Source: nodeSourceFromName(name)})
|
|
return true
|
|
}
|
|
node, err := convertToNode(o)
|
|
if err != nil {
|
|
level.Error(n.logger).Log("msg", "converting to Node object failed", "err", err)
|
|
return true
|
|
}
|
|
send(ctx, ch, n.buildNode(node))
|
|
return true
|
|
}
|
|
|
|
func convertToNode(o interface{}) (*apiv1.Node, error) {
|
|
node, ok := o.(*apiv1.Node)
|
|
if ok {
|
|
return node, nil
|
|
}
|
|
|
|
return nil, errors.Errorf("received unexpected object: %v", o)
|
|
}
|
|
|
|
func nodeSource(n *apiv1.Node) string {
|
|
return nodeSourceFromName(n.Name)
|
|
}
|
|
|
|
func nodeSourceFromName(name string) string {
|
|
return "node/" + name
|
|
}
|
|
|
|
const (
|
|
nodeNameLabel = metaLabelPrefix + "node_name"
|
|
nodeProviderIDLabel = metaLabelPrefix + "node_provider_id"
|
|
nodeLabelPrefix = metaLabelPrefix + "node_label_"
|
|
nodeLabelPresentPrefix = metaLabelPrefix + "node_labelpresent_"
|
|
nodeAnnotationPrefix = metaLabelPrefix + "node_annotation_"
|
|
nodeAnnotationPresentPrefix = metaLabelPrefix + "node_annotationpresent_"
|
|
nodeAddressPrefix = metaLabelPrefix + "node_address_"
|
|
)
|
|
|
|
func nodeLabels(n *apiv1.Node) model.LabelSet {
|
|
// Each label and annotation will create two key-value pairs in the map.
|
|
ls := make(model.LabelSet, 2*(len(n.Labels)+len(n.Annotations))+1)
|
|
|
|
ls[nodeNameLabel] = lv(n.Name)
|
|
ls[nodeProviderIDLabel] = lv(n.Spec.ProviderID)
|
|
|
|
for k, v := range n.Labels {
|
|
ln := strutil.SanitizeLabelName(k)
|
|
ls[model.LabelName(nodeLabelPrefix+ln)] = lv(v)
|
|
ls[model.LabelName(nodeLabelPresentPrefix+ln)] = presentValue
|
|
}
|
|
|
|
for k, v := range n.Annotations {
|
|
ln := strutil.SanitizeLabelName(k)
|
|
ls[model.LabelName(nodeAnnotationPrefix+ln)] = lv(v)
|
|
ls[model.LabelName(nodeAnnotationPresentPrefix+ln)] = presentValue
|
|
}
|
|
return ls
|
|
}
|
|
|
|
func (n *Node) buildNode(node *apiv1.Node) *targetgroup.Group {
|
|
tg := &targetgroup.Group{
|
|
Source: nodeSource(node),
|
|
}
|
|
tg.Labels = nodeLabels(node)
|
|
|
|
addr, addrMap, err := nodeAddress(node)
|
|
if err != nil {
|
|
level.Warn(n.logger).Log("msg", "No node address found", "err", err)
|
|
return nil
|
|
}
|
|
addr = net.JoinHostPort(addr, strconv.FormatInt(int64(node.Status.DaemonEndpoints.KubeletEndpoint.Port), 10))
|
|
|
|
t := model.LabelSet{
|
|
model.AddressLabel: lv(addr),
|
|
model.InstanceLabel: lv(node.Name),
|
|
}
|
|
|
|
for ty, a := range addrMap {
|
|
ln := strutil.SanitizeLabelName(nodeAddressPrefix + string(ty))
|
|
t[model.LabelName(ln)] = lv(a[0])
|
|
}
|
|
tg.Targets = append(tg.Targets, t)
|
|
|
|
return tg
|
|
}
|
|
|
|
// nodeAddresses returns the provided node's address, based on the priority:
|
|
// 1. NodeInternalIP
|
|
// 2. NodeInternalDNS
|
|
// 3. NodeExternalIP
|
|
// 4. NodeExternalDNS
|
|
// 5. NodeLegacyHostIP
|
|
// 6. NodeHostName
|
|
//
|
|
// Derived from k8s.io/kubernetes/pkg/util/node/node.go
|
|
func nodeAddress(node *apiv1.Node) (string, map[apiv1.NodeAddressType][]string, error) {
|
|
m := map[apiv1.NodeAddressType][]string{}
|
|
for _, a := range node.Status.Addresses {
|
|
m[a.Type] = append(m[a.Type], a.Address)
|
|
}
|
|
|
|
if addresses, ok := m[apiv1.NodeInternalIP]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
if addresses, ok := m[apiv1.NodeInternalDNS]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
if addresses, ok := m[apiv1.NodeExternalIP]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
if addresses, ok := m[apiv1.NodeExternalDNS]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
if addresses, ok := m[apiv1.NodeAddressType(NodeLegacyHostIP)]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
if addresses, ok := m[apiv1.NodeHostName]; ok {
|
|
return addresses[0], m, nil
|
|
}
|
|
return "", m, errors.New("host address unknown")
|
|
}
|