mirror of
https://github.com/prometheus/node_exporter.git
synced 2024-12-28 23:19:46 -08:00
successfully exporting one of 2 cpus
This commit is contained in:
parent
cda2f493ee
commit
03da1ea516
|
@ -17,64 +17,113 @@ package collector
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"strconv"
|
"fmt"
|
||||||
"unsafe"
|
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus"
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
)
|
)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
#cgo LDFLAGS:
|
#cgo LDFLAGS:
|
||||||
#include <stdio.h>
|
#include <fcntl.h>
|
||||||
#include <kinfo.h>
|
#include <stdlib.h>
|
||||||
|
#include <sys/param.h>
|
||||||
|
#include <sys/resource.h>
|
||||||
|
#include <sys/time.h>
|
||||||
#include <sys/sysctl.h>
|
#include <sys/sysctl.h>
|
||||||
|
#include <kinfo.h>
|
||||||
|
|
||||||
int
|
static int mibs_set_up = 0;
|
||||||
getCPUTimes(int *ncpu, double **cpu_times, size_t *cp_times_length)
|
|
||||||
{
|
static int mib_kern_cp_times[2];
|
||||||
int cpu, mib[2];
|
static size_t mib_kern_cp_times_len = 2;
|
||||||
struct kinfo_cputime cp_t[ncpu];
|
|
||||||
uint64_t user, nice, sys, intr, idle;
|
static const int mib_hw_ncpu[] = {CTL_HW, HW_NCPU};
|
||||||
|
static const size_t mib_hw_ncpu_len = 2;
|
||||||
|
|
||||||
|
static const int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
|
||||||
|
static size_t mib_kern_clockrate_len = 2;
|
||||||
|
|
||||||
|
// Setup method for MIBs not available as constants.
|
||||||
|
// Calls to this method must be synchronized externally.
|
||||||
|
int setupSysctlMIBs() {
|
||||||
|
int ret = sysctlnametomib("kern.cputime", mib_kern_cp_times, &mib_kern_cp_times_len);
|
||||||
|
if (ret == 0) mibs_set_up = 1;
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
int getCPUTimes(int *ncpu, struct kinfo_cputime *cputime, uint64_t *cpu_user) {
|
||||||
|
// // Assert that mibs are set up through setupSysctlMIBs
|
||||||
|
// if (!mibs_set_up) {
|
||||||
|
// return -1;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// // Retrieve number of cpu cores
|
||||||
|
// size_t ncpu_size = sizeof(*ncpu);
|
||||||
|
// if (sysctl(mib_hw_ncpu, mib_hw_ncpu_len, ncpu, &ncpu_size, NULL, 0) == -1 ||
|
||||||
|
// sizeof(*ncpu) != ncpu_size) {
|
||||||
|
// return -1;
|
||||||
|
// }
|
||||||
|
|
||||||
size_t len;
|
size_t len;
|
||||||
|
|
||||||
|
// Get number of cpu cores.
|
||||||
|
int mib[2];
|
||||||
mib[0] = CTL_HW;
|
mib[0] = CTL_HW;
|
||||||
mib[1] = HW_NCPU;
|
mib[1] = HW_NCPU;
|
||||||
len = sizeof(*ncpu);
|
len = sizeof(*ncpu);
|
||||||
if (-1 == sysctl(mib, 2, &ncpu, &len, NULL, 0))
|
if (sysctl(mib, 2, ncpu, &len, NULL, 0)) {
|
||||||
return -1;
|
|
||||||
|
|
||||||
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
|
|
||||||
|
|
||||||
len = sizeof(cp_t[0])*(*ncpu);
|
|
||||||
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0))
|
|
||||||
return -1;
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
// Retrieve clockrate
|
// Retrieve clockrate
|
||||||
struct clockinfo clockrate;
|
struct clockinfo clockrate;
|
||||||
int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
|
|
||||||
size_t mib_kern_clockrate_len = 2;
|
|
||||||
size_t clockrate_size = sizeof(clockrate);
|
size_t clockrate_size = sizeof(clockrate);
|
||||||
|
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1 ||
|
||||||
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1)
|
sizeof(clockrate) != clockrate_size) {
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
|
|
||||||
cpu_times = (double *) malloc(sizeof(double)*(*cp_times_length));
|
|
||||||
for (int i = 0; i < (*cp_times_length); i++) {
|
|
||||||
(*cpu_times)[i] = ((double) cp_times[i]) / cpufreq;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// // Retrieve cp_times values
|
||||||
|
// *cp_times_length = (*ncpu) * CPUSTATES;
|
||||||
|
//
|
||||||
|
// long cp_times[*cp_times_length];
|
||||||
|
// size_t cp_times_size = sizeof(cp_times);
|
||||||
|
|
||||||
|
// Get the cpu times.
|
||||||
|
struct kinfo_cputime cp_t[*ncpu];
|
||||||
|
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
|
||||||
|
len = sizeof(cp_t[0])*(*ncpu);
|
||||||
|
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0)) {
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
*cpu_user = cp_t[0].cp_user;
|
||||||
|
*cputime = cp_t[0];
|
||||||
|
// This results in outputting:
|
||||||
|
// {1362514572 273421 845667973 12986861 17529717536 0 0 0 0 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]}
|
||||||
|
// So, we have the first 5 numbers from the first cpu.
|
||||||
|
// Need to figure out how to get the second cpu, i.e. cputime needs to be created with [2]
|
||||||
|
|
||||||
|
// Compute absolute time for different CPU states
|
||||||
|
// long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
|
||||||
|
// *cpu_times = (double *) malloc(sizeof(double)*(len));
|
||||||
|
// for (int i = 0; i < (len); i++) {
|
||||||
|
// (*cpu_times)[i] = ((double) cp_t[i]) / cpufreq;
|
||||||
|
// }
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void freeCPUTimes(double *cpu_times) {
|
void freeCPUTimes(double *cpu_times) {
|
||||||
free(cpu_times);
|
free(cpu_times);
|
||||||
}
|
}
|
||||||
*/
|
|
||||||
|
|
||||||
|
*/
|
||||||
import "C"
|
import "C"
|
||||||
|
|
||||||
|
const maxCPUTimesLen = C.MAXCPU * C.CPUSTATES
|
||||||
|
|
||||||
type statCollector struct {
|
type statCollector struct {
|
||||||
cpu *prometheus.CounterVec
|
cpu *prometheus.CounterVec
|
||||||
}
|
}
|
||||||
|
@ -86,6 +135,9 @@ func init() {
|
||||||
// Takes a prometheus registry and returns a new Collector exposing
|
// Takes a prometheus registry and returns a new Collector exposing
|
||||||
// CPU stats.
|
// CPU stats.
|
||||||
func NewStatCollector() (Collector, error) {
|
func NewStatCollector() (Collector, error) {
|
||||||
|
if C.setupSysctlMIBs() == -1 {
|
||||||
|
return nil, errors.New("could not initialize sysctl MIBs")
|
||||||
|
}
|
||||||
return &statCollector{
|
return &statCollector{
|
||||||
cpu: prometheus.NewCounterVec(
|
cpu: prometheus.NewCounterVec(
|
||||||
prometheus.CounterOpts{
|
prometheus.CounterOpts{
|
||||||
|
@ -98,30 +150,50 @@ func NewStatCollector() (Collector, error) {
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type kinfoCPUTime struct {
|
||||||
|
cp_user, cp_nice, cp_sys, cp_intr, cp_idle uint64
|
||||||
|
}
|
||||||
|
|
||||||
// Expose CPU stats using sysctl.
|
// Expose CPU stats using sysctl.
|
||||||
func (c *statCollector) Update(ch chan<- prometheus.Metric) (err error) {
|
func (c *statCollector) Update(ch chan<- prometheus.Metric) (err error) {
|
||||||
// Adapted from
|
|
||||||
// https://www.dragonflybsd.org/mailarchive/users/2010-04/msg00056.html
|
// We want time spent per-cpu per CPUSTATE.
|
||||||
|
// CPUSTATES (number of CPUSTATES) is defined as 5U.
|
||||||
|
// Order: CP_USER | CP_NICE | CP_SYS | CP_IDLE | CP_INTR
|
||||||
|
// sysctl kern.cp_times provides hw.ncpu * CPUSTATES long integers:
|
||||||
|
// hw.ncpu * (space-separated list of the above variables)
|
||||||
|
//
|
||||||
|
// Each value is a counter incremented at frequency
|
||||||
|
// kern.clockrate.(stathz | hz)
|
||||||
|
//
|
||||||
|
// Look into sys/kern/kern_clock.c for details.
|
||||||
|
|
||||||
var ncpu C.int
|
var ncpu C.int
|
||||||
var cpuTimesC *C.double
|
var cpuTimesC C.struct_kinfo_cputime
|
||||||
var cpuTimesLength C.size_t
|
var cpuTimesLength C.uint64_t
|
||||||
|
|
||||||
if C.getCPUTimes(&ncpu, &cpuTimesC, &cpuTimesLength) == -1 {
|
if C.getCPUTimes(&ncpu, &cpuTimesC, &cpuTimesLength) == -1 {
|
||||||
return errors.New("could not retrieve CPU times")
|
return errors.New("could not retrieve CPU times")
|
||||||
}
|
}
|
||||||
defer C.freeCPUTimes(cpuTimesC)
|
// defer C.freeCPUTimes(cpuTimesC)
|
||||||
|
fmt.Println(cpuTimesC)
|
||||||
|
fmt.Println(uint64(cpuTimesLength))
|
||||||
|
return errors.New("early kill")
|
||||||
|
if cpuTimesLength > maxCPUTimesLen {
|
||||||
|
return errors.New("more CPU's than MAXCPU?")
|
||||||
|
}
|
||||||
|
|
||||||
// Convert C.double array to Go array (https://github.com/golang/go/wiki/cgo#turning-c-arrays-into-go-slices).
|
// Convert C.double array to Go array (https://github.com/golang/go/wiki/cgo#turning-c-arrays-into-go-slices).
|
||||||
cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
|
// cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
|
||||||
|
//
|
||||||
for cpu := 0; cpu < int(ncpu); cpu++ {
|
// for cpu := 0; cpu < int(ncpu); cpu++ {
|
||||||
base_idx := C.CPUSTATES * cpu
|
// base_idx := C.CPUSTATES * cpu
|
||||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
|
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
|
||||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
|
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
|
||||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
|
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
|
||||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
|
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
|
||||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
|
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
|
||||||
}
|
// }
|
||||||
|
|
||||||
c.cpu.Collect(ch)
|
c.cpu.Collect(ch)
|
||||||
return err
|
return err
|
||||||
|
|
Loading…
Reference in a new issue