perf tools: Add ordered_events__(new|delete) interface

Adding new ordered events interface to new|delete event buffer:

  ordered_events__new    - allocate event buffer from the cache
  ordered_events__delete - return event buffer to the cache

Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Acked-by: David Ahern <dsahern@gmail.com>
Cc: Corey Ashford <cjashfor@linux.vnet.ibm.com>
Cc: David Ahern <dsahern@gmail.com>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jean Pihet <jean.pihet@linaro.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Link: http://lkml.kernel.org/n/tip-srwunsy7o5wl17vpt4a10oxp@git.kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
Jiri Olsa 2014-06-10 21:58:02 +02:00 committed by Arnaldo Carvalho de Melo
parent fc12482f4f
commit c64c7e1a5a

View file

@ -464,6 +464,100 @@ static void perf_session_free_sample_buffers(struct perf_session *session)
}
}
/* The queue is ordered by time */
static void queue_event(struct ordered_events *oe, struct ordered_event *new)
{
struct ordered_event *last = oe->last;
u64 timestamp = new->timestamp;
struct list_head *p;
++oe->nr_events;
oe->last = new;
if (!last) {
list_add(&new->list, &oe->events);
oe->max_timestamp = timestamp;
return;
}
/*
* last event might point to some random place in the list as it's
* the last queued event. We expect that the new event is close to
* this.
*/
if (last->timestamp <= timestamp) {
while (last->timestamp <= timestamp) {
p = last->list.next;
if (p == &oe->events) {
list_add_tail(&new->list, &oe->events);
oe->max_timestamp = timestamp;
return;
}
last = list_entry(p, struct ordered_event, list);
}
list_add_tail(&new->list, &last->list);
} else {
while (last->timestamp > timestamp) {
p = last->list.prev;
if (p == &oe->events) {
list_add(&new->list, &oe->events);
return;
}
last = list_entry(p, struct ordered_event, list);
}
list_add(&new->list, &last->list);
}
}
#define MAX_SAMPLE_BUFFER (64 * 1024 / sizeof(struct ordered_event))
static struct ordered_event *alloc_event(struct ordered_events *oe)
{
struct list_head *cache = &oe->cache;
struct ordered_event *new;
if (!list_empty(cache)) {
new = list_entry(cache->next, struct ordered_event, list);
list_del(&new->list);
} else if (oe->buffer) {
new = oe->buffer + oe->buffer_idx;
if (++oe->buffer_idx == MAX_SAMPLE_BUFFER)
oe->buffer = NULL;
} else {
oe->buffer = malloc(MAX_SAMPLE_BUFFER * sizeof(*new));
if (!oe->buffer)
return NULL;
list_add(&oe->buffer->list, &oe->to_free);
/* First entry is abused to maintain the to_free list. */
oe->buffer_idx = 2;
new = oe->buffer + 1;
}
return new;
}
static struct ordered_event *
ordered_events__new(struct ordered_events *oe, u64 timestamp)
{
struct ordered_event *new;
new = alloc_event(oe);
if (new) {
new->timestamp = timestamp;
queue_event(oe, new);
}
return new;
}
static void
ordered_events__delete(struct ordered_events *oe, struct ordered_event *event)
{
list_del(&event->list);
list_add(&event->list, &oe->cache);
oe->nr_events--;
}
static int perf_session_deliver_event(struct perf_session *session,
union perf_event *event,
struct perf_sample *sample,
@ -506,10 +600,8 @@ static int ordered_events__flush(struct perf_session *s,
return ret;
}
ordered_events__delete(oe, iter);
oe->last_flush = iter->timestamp;
list_del(&iter->list);
list_add(&iter->list, &oe->cache);
oe->nr_events--;
if (show_progress)
ui_progress__update(&prog, 1);
@ -573,59 +665,10 @@ static int process_finished_round(struct perf_tool *tool,
return ret;
}
/* The queue is ordered by time */
static void __queue_event(struct ordered_event *new, struct perf_session *s)
{
struct ordered_events *oe = &s->ordered_events;
struct ordered_event *last = oe->last;
u64 timestamp = new->timestamp;
struct list_head *p;
++oe->nr_events;
oe->last = new;
if (!last) {
list_add(&new->list, &oe->events);
oe->max_timestamp = timestamp;
return;
}
/*
* last event might point to some random place in the list as it's
* the last queued event. We expect that the new event is close to
* this.
*/
if (last->timestamp <= timestamp) {
while (last->timestamp <= timestamp) {
p = last->list.next;
if (p == &oe->events) {
list_add_tail(&new->list, &oe->events);
oe->max_timestamp = timestamp;
return;
}
last = list_entry(p, struct ordered_event, list);
}
list_add_tail(&new->list, &last->list);
} else {
while (last->timestamp > timestamp) {
p = last->list.prev;
if (p == &oe->events) {
list_add(&new->list, &oe->events);
return;
}
last = list_entry(p, struct ordered_event, list);
}
list_add(&new->list, &last->list);
}
}
#define MAX_SAMPLE_BUFFER (64 * 1024 / sizeof(struct ordered_event))
int perf_session_queue_event(struct perf_session *s, union perf_event *event,
struct perf_sample *sample, u64 file_offset)
{
struct ordered_events *oe = &s->ordered_events;
struct list_head *cache = &oe->cache;
u64 timestamp = sample->time;
struct ordered_event *new;
@ -637,28 +680,12 @@ int perf_session_queue_event(struct perf_session *s, union perf_event *event,
return -EINVAL;
}
if (!list_empty(cache)) {
new = list_entry(cache->next, struct ordered_event, list);
list_del(&new->list);
} else if (oe->buffer) {
new = oe->buffer + oe->buffer_idx;
if (++oe->buffer_idx == MAX_SAMPLE_BUFFER)
oe->buffer = NULL;
} else {
oe->buffer = malloc(MAX_SAMPLE_BUFFER * sizeof(*new));
if (!oe->buffer)
return -ENOMEM;
list_add(&oe->buffer->list, &oe->to_free);
oe->buffer_idx = 2;
new = oe->buffer + 1;
}
new = ordered_events__new(oe, timestamp);
if (!new)
return -ENOMEM;
new->timestamp = timestamp;
new->file_offset = file_offset;
new->event = event;
__queue_event(new, s);
return 0;
}