blob: 3ae2266f948bd8a747064b6f29615d325343df5a [file] [log] [blame]
/*
* (C) Copyright 2009-2011 -
* Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
*
* LTTng performance monitoring counters (perf-counters) integration module.
*
* Dual LGPL v2.1/GPL v2 license.
*/
#include <linux/module.h>
#include <linux/slab.h>
#include <linux/perf_event.h>
#include <linux/list.h>
#include <linux/string.h>
#include "ltt-events.h"
#include "wrapper/ringbuffer/frontend_types.h"
#include "wrapper/vmalloc.h"
#include "wrapper/perf.h"
#include "ltt-tracer.h"
static
size_t perf_counter_get_size(size_t offset)
{
size_t size = 0;
size += lib_ring_buffer_align(offset, ltt_alignof(uint64_t));
size += sizeof(uint64_t);
return size;
}
static
void perf_counter_record(struct lttng_ctx_field *field,
struct lib_ring_buffer_ctx *ctx,
struct ltt_channel *chan)
{
struct perf_event *event;
uint64_t value;
event = field->u.perf_counter->e[ctx->cpu];
if (likely(event)) {
if (unlikely(event->state == PERF_EVENT_STATE_ERROR)) {
value = 0;
} else {
event->pmu->read(event);
value = local64_read(&event->count);
}
} else {
/*
* Perf chooses not to be clever and not to support enabling a
* perf counter before the cpu is brought up. Therefore, we need
* to support having events coming (e.g. scheduler events)
* before the counter is setup. Write an arbitrary 0 in this
* case.
*/
value = 0;
}
lib_ring_buffer_align_ctx(ctx, ltt_alignof(value));
chan->ops->event_write(ctx, &value, sizeof(value));
}
#if defined(CONFIG_PERF_EVENTS) && (LINUX_VERSION_CODE >= KERNEL_VERSION(3,0,99))
static
void overflow_callback(struct perf_event *event,
struct perf_sample_data *data,
struct pt_regs *regs)
{
}
#else
static
void overflow_callback(struct perf_event *event, int nmi,
struct perf_sample_data *data,
struct pt_regs *regs)
{
}
#endif
static
void lttng_destroy_perf_counter_field(struct lttng_ctx_field *field)
{
struct perf_event **events = field->u.perf_counter->e;
int cpu;
get_online_cpus();
for_each_online_cpu(cpu)
perf_event_release_kernel(events[cpu]);
put_online_cpus();
#ifdef CONFIG_HOTPLUG_CPU
unregister_cpu_notifier(&field->u.perf_counter->nb);
#endif
kfree(field->event_field.name);
kfree(field->u.perf_counter->attr);
kfree(events);
kfree(field->u.perf_counter);
}
#ifdef CONFIG_HOTPLUG_CPU
/**
* lttng_perf_counter_hp_callback - CPU hotplug callback
* @nb: notifier block
* @action: hotplug action to take
* @hcpu: CPU number
*
* Returns the success/failure of the operation. (%NOTIFY_OK, %NOTIFY_BAD)
*
* We can setup perf counters when the cpu is online (up prepare seems to be too
* soon).
*/
static
int __cpuinit lttng_perf_counter_cpu_hp_callback(struct notifier_block *nb,
unsigned long action,
void *hcpu)
{
unsigned int cpu = (unsigned long) hcpu;
struct lttng_perf_counter_field *perf_field =
container_of(nb, struct lttng_perf_counter_field, nb);
struct perf_event **events = perf_field->e;
struct perf_event_attr *attr = perf_field->attr;
struct perf_event *pevent;
if (!perf_field->hp_enable)
return NOTIFY_OK;
switch (action) {
case CPU_ONLINE:
case CPU_ONLINE_FROZEN:
pevent = wrapper_perf_event_create_kernel_counter(attr,
cpu, NULL, overflow_callback);
if (!pevent || IS_ERR(pevent))
return NOTIFY_BAD;
if (pevent->state == PERF_EVENT_STATE_ERROR) {
perf_event_release_kernel(pevent);
return NOTIFY_BAD;
}
barrier(); /* Create perf counter before setting event */
events[cpu] = pevent;
break;
case CPU_UP_CANCELED:
case CPU_UP_CANCELED_FROZEN:
case CPU_DEAD:
case CPU_DEAD_FROZEN:
pevent = events[cpu];
events[cpu] = NULL;
barrier(); /* NULLify event before perf counter teardown */
perf_event_release_kernel(pevent);
break;
}
return NOTIFY_OK;
}
#endif
int lttng_add_perf_counter_to_ctx(uint32_t type,
uint64_t config,
const char *name,
struct lttng_ctx **ctx)
{
struct lttng_ctx_field *field;
struct lttng_perf_counter_field *perf_field;
struct perf_event **events;
struct perf_event_attr *attr;
int ret;
int cpu;
char *name_alloc;
events = kzalloc(num_possible_cpus() * sizeof(*events), GFP_KERNEL);
if (!events)
return -ENOMEM;
attr = kzalloc(sizeof(struct perf_event_attr), GFP_KERNEL);
if (!attr) {
ret = -ENOMEM;
goto error_attr;
}
attr->type = type;
attr->config = config;
attr->size = sizeof(struct perf_event_attr);
attr->pinned = 1;
attr->disabled = 0;
perf_field = kzalloc(sizeof(struct lttng_perf_counter_field), GFP_KERNEL);
if (!perf_field) {
ret = -ENOMEM;
goto error_alloc_perf_field;
}
perf_field->e = events;
perf_field->attr = attr;
name_alloc = kstrdup(name, GFP_KERNEL);
if (!name_alloc) {
ret = -ENOMEM;
goto name_alloc_error;
}
field = lttng_append_context(ctx);
if (!field) {
ret = -ENOMEM;
goto append_context_error;
}
if (lttng_find_context(*ctx, name_alloc)) {
ret = -EEXIST;
goto find_error;
}
#ifdef CONFIG_HOTPLUG_CPU
perf_field->nb.notifier_call =
lttng_perf_counter_cpu_hp_callback;
perf_field->nb.priority = 0;
register_cpu_notifier(&perf_field->nb);
#endif
get_online_cpus();
for_each_online_cpu(cpu) {
events[cpu] = wrapper_perf_event_create_kernel_counter(attr,
cpu, NULL, overflow_callback);
if (!events[cpu] || IS_ERR(events[cpu])) {
ret = -EINVAL;
goto counter_error;
}
if (events[cpu]->state == PERF_EVENT_STATE_ERROR) {
ret = -EBUSY;
goto counter_busy;
}
}
put_online_cpus();
field->destroy = lttng_destroy_perf_counter_field;
field->event_field.name = name_alloc;
field->event_field.type.atype = atype_integer;
field->event_field.type.u.basic.integer.size = sizeof(uint64_t) * CHAR_BIT;
field->event_field.type.u.basic.integer.alignment = ltt_alignof(uint64_t) * CHAR_BIT;
field->event_field.type.u.basic.integer.signedness = is_signed_type(uint64_t);
field->event_field.type.u.basic.integer.reverse_byte_order = 0;
field->event_field.type.u.basic.integer.base = 10;
field->event_field.type.u.basic.integer.encoding = lttng_encode_none;
field->get_size = perf_counter_get_size;
field->record = perf_counter_record;
field->u.perf_counter = perf_field;
perf_field->hp_enable = 1;
wrapper_vmalloc_sync_all();
return 0;
counter_busy:
counter_error:
for_each_online_cpu(cpu) {
if (events[cpu] && !IS_ERR(events[cpu]))
perf_event_release_kernel(events[cpu]);
}
put_online_cpus();
#ifdef CONFIG_HOTPLUG_CPU
unregister_cpu_notifier(&perf_field->nb);
#endif
find_error:
lttng_remove_context_field(ctx, field);
append_context_error:
kfree(name_alloc);
name_alloc_error:
kfree(perf_field);
error_alloc_perf_field:
kfree(attr);
error_attr:
kfree(events);
return ret;
}
MODULE_LICENSE("GPL and additional rights");
MODULE_AUTHOR("Mathieu Desnoyers");
MODULE_DESCRIPTION("Linux Trace Toolkit Perf Support");