diff --git a/collector/cpu_dragonfly.go b/collector/cpu_dragonfly.go index fd36418f..0d414617 100644 --- a/collector/cpu_dragonfly.go +++ b/collector/cpu_dragonfly.go @@ -17,64 +17,113 @@ package collector import ( "errors" - "strconv" - "unsafe" + "fmt" "github.com/prometheus/client_golang/prometheus" ) /* #cgo LDFLAGS: -#include -#include +#include +#include +#include +#include +#include #include +#include -int -getCPUTimes(int *ncpu, double **cpu_times, size_t *cp_times_length) -{ - int cpu, mib[2]; - struct kinfo_cputime cp_t[ncpu]; - uint64_t user, nice, sys, intr, idle; +static int mibs_set_up = 0; + +static int mib_kern_cp_times[2]; +static size_t mib_kern_cp_times_len = 2; + +static const int mib_hw_ncpu[] = {CTL_HW, HW_NCPU}; +static const size_t mib_hw_ncpu_len = 2; + +static const int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE}; +static size_t mib_kern_clockrate_len = 2; + +// Setup method for MIBs not available as constants. +// Calls to this method must be synchronized externally. +int setupSysctlMIBs() { + int ret = sysctlnametomib("kern.cputime", mib_kern_cp_times, &mib_kern_cp_times_len); + if (ret == 0) mibs_set_up = 1; + return ret; +} + +int getCPUTimes(int *ncpu, struct kinfo_cputime *cputime, uint64_t *cpu_user) { + // // Assert that mibs are set up through setupSysctlMIBs + // if (!mibs_set_up) { + // return -1; + // } + + // // Retrieve number of cpu cores + // size_t ncpu_size = sizeof(*ncpu); + // if (sysctl(mib_hw_ncpu, mib_hw_ncpu_len, ncpu, &ncpu_size, NULL, 0) == -1 || + // sizeof(*ncpu) != ncpu_size) { + // return -1; + // } size_t len; + // Get number of cpu cores. + int mib[2]; mib[0] = CTL_HW; mib[1] = HW_NCPU; len = sizeof(*ncpu); - if (-1 == sysctl(mib, 2, &ncpu, &len, NULL, 0)) - return -1; - - bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu)); - - len = sizeof(cp_t[0])*(*ncpu); - if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0)) + if (sysctl(mib, 2, ncpu, &len, NULL, 0)) { return -1; + } // Retrieve clockrate struct clockinfo clockrate; - int mib_kern_clockrate[] = {CTL_KERN, KERN_CLOCKRATE}; - size_t mib_kern_clockrate_len = 2; size_t clockrate_size = sizeof(clockrate); - - if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1) + if (sysctl(mib_kern_clockrate, mib_kern_clockrate_len, &clockrate, &clockrate_size, NULL, 0) == -1 || + sizeof(clockrate) != clockrate_size) { return -1; - - long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz; - cpu_times = (double *) malloc(sizeof(double)*(*cp_times_length)); - for (int i = 0; i < (*cp_times_length); i++) { - (*cpu_times)[i] = ((double) cp_times[i]) / cpufreq; } + // // Retrieve cp_times values + // *cp_times_length = (*ncpu) * CPUSTATES; + // + // long cp_times[*cp_times_length]; + // size_t cp_times_size = sizeof(cp_times); + + // Get the cpu times. + struct kinfo_cputime cp_t[*ncpu]; + bzero(cp_t, sizeof(struct kinfo_cputime)*(*ncpu)); + len = sizeof(cp_t[0])*(*ncpu); + if (sysctlbyname("kern.cputime", &cp_t, &len, NULL, 0)) { + return -1; + } + + *cpu_user = cp_t[0].cp_user; + *cputime = cp_t[0]; + // This results in outputting: + // {1362514572 273421 845667973 12986861 17529717536 0 0 0 0 [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]} + // So, we have the first 5 numbers from the first cpu. + // Need to figure out how to get the second cpu, i.e. cputime needs to be created with [2] + + // Compute absolute time for different CPU states + // long cpufreq = clockrate.stathz > 0 ? clockrate.stathz : clockrate.hz; + // *cpu_times = (double *) malloc(sizeof(double)*(len)); + // for (int i = 0; i < (len); i++) { + // (*cpu_times)[i] = ((double) cp_t[i]) / cpufreq; + // } + return 0; + } void freeCPUTimes(double *cpu_times) { free(cpu_times); } -*/ +*/ import "C" +const maxCPUTimesLen = C.MAXCPU * C.CPUSTATES + type statCollector struct { cpu *prometheus.CounterVec } @@ -86,6 +135,9 @@ func init() { // Takes a prometheus registry and returns a new Collector exposing // CPU stats. func NewStatCollector() (Collector, error) { + if C.setupSysctlMIBs() == -1 { + return nil, errors.New("could not initialize sysctl MIBs") + } return &statCollector{ cpu: prometheus.NewCounterVec( prometheus.CounterOpts{ @@ -98,30 +150,50 @@ func NewStatCollector() (Collector, error) { }, nil } +type kinfoCPUTime struct { + cp_user, cp_nice, cp_sys, cp_intr, cp_idle uint64 +} + // Expose CPU stats using sysctl. func (c *statCollector) Update(ch chan<- prometheus.Metric) (err error) { - // Adapted from - // https://www.dragonflybsd.org/mailarchive/users/2010-04/msg00056.html + + // We want time spent per-cpu per CPUSTATE. + // CPUSTATES (number of CPUSTATES) is defined as 5U. + // Order: CP_USER | CP_NICE | CP_SYS | CP_IDLE | CP_INTR + // sysctl kern.cp_times provides hw.ncpu * CPUSTATES long integers: + // hw.ncpu * (space-separated list of the above variables) + // + // Each value is a counter incremented at frequency + // kern.clockrate.(stathz | hz) + // + // Look into sys/kern/kern_clock.c for details. var ncpu C.int - var cpuTimesC *C.double - var cpuTimesLength C.size_t + var cpuTimesC C.struct_kinfo_cputime + var cpuTimesLength C.uint64_t + if C.getCPUTimes(&ncpu, &cpuTimesC, &cpuTimesLength) == -1 { return errors.New("could not retrieve CPU times") } - defer C.freeCPUTimes(cpuTimesC) + // defer C.freeCPUTimes(cpuTimesC) + fmt.Println(cpuTimesC) + fmt.Println(uint64(cpuTimesLength)) + return errors.New("early kill") + if cpuTimesLength > maxCPUTimesLen { + return errors.New("more CPU's than MAXCPU?") + } // Convert C.double array to Go array (https://github.com/golang/go/wiki/cgo#turning-c-arrays-into-go-slices). - cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength] - - for cpu := 0; cpu < int(ncpu); cpu++ { - base_idx := C.CPUSTATES * cpu - c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER])) - c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE])) - c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS])) - c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR])) - c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE])) - } + // cpuTimes := (*[maxCPUTimesLen]C.double)(unsafe.Pointer(cpuTimesC))[:cpuTimesLength:cpuTimesLength] + // + // for cpu := 0; cpu < int(ncpu); cpu++ { + // base_idx := C.CPUSTATES * cpu + // c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "user"}).Set(float64(cpuTimes[base_idx+C.CP_USER])) + // c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "nice"}).Set(float64(cpuTimes[base_idx+C.CP_NICE])) + // c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "system"}).Set(float64(cpuTimes[base_idx+C.CP_SYS])) + // c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "interrupt"}).Set(float64(cpuTimes[base_idx+C.CP_INTR])) + // c.cpu.With(prometheus.Labels{"cpu": strconv.Itoa(cpu), "mode": "idle"}).Set(float64(cpuTimes[base_idx+C.CP_IDLE])) + // } c.cpu.Collect(ch) return err