mirror of
https://github.com/prometheus/node_exporter.git
synced 2024-12-26 22:19:44 -08:00
successfully exporting one of 2 cpus
This commit is contained in:
parent
cda2f493ee
commit
03da1ea516
|
@ -17,64 +17,113 @@ package collector
|
|||
|
||||
import (
|
||||
"errors"
|
||||
"strconv"
|
||||
"unsafe"
|
||||
"fmt"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
/*
|
||||
#cgo LDFLAGS:
|
||||
#include <stdio.h>
|
||||
#include <kinfo.h>
|
||||
#include <fcntl.h>
|
||||
#include <stdlib.h>
|
||||
#include <sys/param.h>
|
||||
#include <sys/resource.h>
|
||||
#include <sys/time.h>
|
||||
#include <sys/sysctl.h>
|
||||
#include <kinfo.h>
|
||||
|
||||
int
|
||||
getCPUTimes(int *ncpu, double **cpu_times, size_t *cp_times_length)
|
||||
{
|
||||
int cpu, mib[2];
|
||||
struct kinfo_cputime cp_t[ncpu];
|
||||
uint64_t user, nice, sys, intr, idle;
|
||||
static int mibs_set_up = 0;
|
||||
|
||||
static int mib_kern_cp_times[2];
|
||||
static size_t mib_kern_cp_times_len = 2;
|
||||
|
||||
static const int mib_hw_ncpu[] = {CTL_HW, HW_NCPU};
|
||||
static const size_t mib_hw_ncpu_len = 2;
|
||||
|
||||
static const int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
|
||||
static size_t mib_kern_clockrate_len = 2;
|
||||
|
||||
// Setup method for MIBs not available as constants.
|
||||
// Calls to this method must be synchronized externally.
|
||||
int setupSysctlMIBs() {
|
||||
int ret = sysctlnametomib("kern.cputime", mib_kern_cp_times, &mib_kern_cp_times_len);
|
||||
if (ret == 0) mibs_set_up = 1;
|
||||
return ret;
|
||||
}
|
||||
|
||||
int getCPUTimes(int *ncpu, struct kinfo_cputime *cputime, uint64_t *cpu_user) {
|
||||
// // Assert that mibs are set up through setupSysctlMIBs
|
||||
// if (!mibs_set_up) {
|
||||
// return -1;
|
||||
// }
|
||||
|
||||
// // Retrieve number of cpu cores
|
||||
// size_t ncpu_size = sizeof(*ncpu);
|
||||
// if (sysctl(mib_hw_ncpu, mib_hw_ncpu_len, ncpu, &ncpu_size, NULL, 0) == -1 ||
|
||||
// sizeof(*ncpu) != ncpu_size) {
|
||||
// return -1;
|
||||
// }
|
||||
|
||||
size_t len;
|
||||
|
||||
// Get number of cpu cores.
|
||||
int mib[2];
|
||||
mib[0] = CTL_HW;
|
||||
mib[1] = HW_NCPU;
|
||||
len = sizeof(*ncpu);
|
||||
if (-1 == sysctl(mib, 2, &ncpu, &len, NULL, 0))
|
||||
return -1;
|
||||
|
||||
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
|
||||
|
||||
len = sizeof(cp_t[0])*(*ncpu);
|
||||
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0))
|
||||
if (sysctl(mib, 2, ncpu, &len, NULL, 0)) {
|
||||
return -1;
|
||||
}
|
||||
|
||||
// Retrieve clockrate
|
||||
struct clockinfo clockrate;
|
||||
int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE};
|
||||
size_t mib_kern_clockrate_len = 2;
|
||||
size_t clockrate_size = sizeof(clockrate);
|
||||
|
||||
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1)
|
||||
if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1 ||
|
||||
sizeof(clockrate) != clockrate_size) {
|
||||
return -1;
|
||||
|
||||
long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
|
||||
cpu_times = (double *) malloc(sizeof(double)*(*cp_times_length));
|
||||
for (int i = 0; i < (*cp_times_length); i++) {
|
||||
(*cpu_times)[i] = ((double) cp_times[i]) / cpufreq;
|
||||
}
|
||||
|
||||
// // Retrieve cp_times values
|
||||
// *cp_times_length = (*ncpu) * CPUSTATES;
|
||||
//
|
||||
// long cp_times[*cp_times_length];
|
||||
// size_t cp_times_size = sizeof(cp_times);
|
||||
|
||||
// Get the cpu times.
|
||||
struct kinfo_cputime cp_t[*ncpu];
|
||||
bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu));
|
||||
len = sizeof(cp_t[0])*(*ncpu);
|
||||
if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0)) {
|
||||
return -1;
|
||||
}
|
||||
|
||||
*cpu_user = cp_t[0].cp_user;
|
||||
*cputime = cp_t[0];
|
||||
// This results in outputting:
|
||||
// {1362514572 273421 845667973 12986861 17529717536 0 0 0 0 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]}
|
||||
// So, we have the first 5 numbers from the first cpu.
|
||||
// Need to figure out how to get the second cpu, i.e. cputime needs to be created with [2]
|
||||
|
||||
// Compute absolute time for different CPU states
|
||||
// long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz;
|
||||
// *cpu_times = (double *) malloc(sizeof(double)*(len));
|
||||
// for (int i = 0; i < (len); i++) {
|
||||
// (*cpu_times)[i] = ((double) cp_t[i]) / cpufreq;
|
||||
// }
|
||||
|
||||
return 0;
|
||||
|
||||
}
|
||||
|
||||
void freeCPUTimes(double *cpu_times) {
|
||||
free(cpu_times);
|
||||
}
|
||||
*/
|
||||
|
||||
*/
|
||||
import "C"
|
||||
|
||||
const maxCPUTimesLen = C.MAXCPU * C.CPUSTATES
|
||||
|
||||
type statCollector struct {
|
||||
cpu *prometheus.CounterVec
|
||||
}
|
||||
|
@ -86,6 +135,9 @@ func init() {
|
|||
// Takes a prometheus registry and returns a new Collector exposing
|
||||
// CPU stats.
|
||||
func NewStatCollector() (Collector, error) {
|
||||
if C.setupSysctlMIBs() == -1 {
|
||||
return nil, errors.New("could not initialize sysctl MIBs")
|
||||
}
|
||||
return &statCollector{
|
||||
cpu: prometheus.NewCounterVec(
|
||||
prometheus.CounterOpts{
|
||||
|
@ -98,30 +150,50 @@ func NewStatCollector() (Collector, error) {
|
|||
}, nil
|
||||
}
|
||||
|
||||
type kinfoCPUTime struct {
|
||||
cp_user, cp_nice, cp_sys, cp_intr, cp_idle uint64
|
||||
}
|
||||
|
||||
// Expose CPU stats using sysctl.
|
||||
func (c *statCollector) Update(ch chan<- prometheus.Metric) (err error) {
|
||||
// Adapted from
|
||||
// https://www.dragonflybsd.org/mailarchive/users/2010-04/msg00056.html
|
||||
|
||||
// We want time spent per-cpu per CPUSTATE.
|
||||
// CPUSTATES (number of CPUSTATES) is defined as 5U.
|
||||
// Order: CP_USER | CP_NICE | CP_SYS | CP_IDLE | CP_INTR
|
||||
// sysctl kern.cp_times provides hw.ncpu * CPUSTATES long integers:
|
||||
// hw.ncpu * (space-separated list of the above variables)
|
||||
//
|
||||
// Each value is a counter incremented at frequency
|
||||
// kern.clockrate.(stathz | hz)
|
||||
//
|
||||
// Look into sys/kern/kern_clock.c for details.
|
||||
|
||||
var ncpu C.int
|
||||
var cpuTimesC *C.double
|
||||
var cpuTimesLength C.size_t
|
||||
var cpuTimesC C.struct_kinfo_cputime
|
||||
var cpuTimesLength C.uint64_t
|
||||
|
||||
if C.getCPUTimes(&ncpu, &cpuTimesC, &cpuTimesLength) == -1 {
|
||||
return errors.New("could not retrieve CPU times")
|
||||
}
|
||||
defer C.freeCPUTimes(cpuTimesC)
|
||||
// defer C.freeCPUTimes(cpuTimesC)
|
||||
fmt.Println(cpuTimesC)
|
||||
fmt.Println(uint64(cpuTimesLength))
|
||||
return errors.New("early kill")
|
||||
if cpuTimesLength > maxCPUTimesLen {
|
||||
return errors.New("more CPU's than MAXCPU?")
|
||||
}
|
||||
|
||||
// Convert C.double array to Go array (https://github.com/golang/go/wiki/cgo#turning-c-arrays-into-go-slices).
|
||||
cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
|
||||
|
||||
for cpu := 0; cpu < int(ncpu); cpu++ {
|
||||
base_idx := C.CPUSTATES * cpu
|
||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
|
||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
|
||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
|
||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
|
||||
c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
|
||||
}
|
||||
// cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength]
|
||||
//
|
||||
// for cpu := 0; cpu < int(ncpu); cpu++ {
|
||||
// base_idx := C.CPUSTATES * cpu
|
||||
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER]))
|
||||
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE]))
|
||||
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS]))
|
||||
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR]))
|
||||
// c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE]))
|
||||
// }
|
||||
|
||||
c.cpu.Collect(ch)
|
||||
return err
|
||||
|
|
Loading…
Reference in a new issue