ntel-gpu-tools/overlay/gpu-perf.c
Chris Wilson 6233cac9c9 overlay: Couple wait begin/end events together to fix accounting
Since the events may be processed out of order (due to per-cpu
ringbuffers) we need to be careful to associated wait pairs in order to
compute the correct elapsed time.

Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
2013-08-17 22:22:21 +01:00

382 lines
7.9 KiB
C

#include <stdint.h>
#include <stdbool.h>
#include <linux/perf_event.h>
#include <sys/types.h>
#include <sys/mman.h>
#include <sys/ioctl.h>
#include <unistd.h>
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <fcntl.h>
#include <errno.h>
#include "gpu-perf.h"
#if defined(__i386__)
#define rmb() asm volatile("lock; addl $0,0(%%esp)" ::: "memory")
#endif
#if defined(__x86_64__)
#define rmb() asm volatile("lfence" ::: "memory")
#endif
#define TRACING_EVENT_PATH "/sys/kernel/debug/tracing/events"
#define N_PAGES 32
struct sample_event {
struct perf_event_header header;
uint32_t pid, tid;
uint64_t time;
uint64_t id;
uint32_t raw_size;
uint32_t raw_hdr0;
uint32_t raw_hdr1;
uint32_t raw[0];
};
static int
perf_event_open(struct perf_event_attr *attr,
pid_t pid,
int cpu,
int group_fd,
unsigned long flags)
{
#ifndef __NR_perf_event_open
#if defined(__i386__)
#define __NR_perf_event_open 336
#elif defined(__x86_64__)
#define __NR_perf_event_open 298
#else
#define __NR_perf_event_open 0
#endif
#endif
attr->size = sizeof(*attr);
return syscall(__NR_perf_event_open, attr, pid, cpu, group_fd, flags);
}
static uint64_t tracepoint_id(const char *sys, const char *name)
{
char buf[1024];
int fd, n;
snprintf(buf, sizeof(buf), "%s/%s/%s/id", TRACING_EVENT_PATH, sys, name);
fd = open(buf, 0);
if (fd < 0)
return 0;
n = read(fd, buf, sizeof(buf)-1);
close(fd);
if (n < 0)
return 0;
buf[n] = '\0';
return strtoull(buf, 0, 0);
}
static int perf_tracepoint_open(struct gpu_perf *gp,
const char *sys, const char *name,
int (*func)(struct gpu_perf *, const void *))
{
struct perf_event_attr attr;
struct gpu_perf_sample *sample;
int n, *fd;
memset(&attr, 0, sizeof (attr));
attr.type = PERF_TYPE_TRACEPOINT;
attr.config = tracepoint_id(sys, name);
if (attr.config == 0)
return ENOENT;
attr.sample_period = 1;
attr.sample_type = (PERF_SAMPLE_TIME | PERF_SAMPLE_STREAM_ID | PERF_SAMPLE_TID | PERF_SAMPLE_RAW);
attr.read_format = PERF_FORMAT_ID;
attr.exclude_guest = 1;
n = gp->nr_cpus * (gp->nr_events+1);
fd = realloc(gp->fd, n*sizeof(int));
sample = realloc(gp->sample, n*sizeof(*gp->sample));
if (fd == NULL || sample == NULL)
return ENOMEM;
gp->fd = fd;
gp->sample = sample;
fd += gp->nr_events * gp->nr_cpus;
sample += gp->nr_events * gp->nr_cpus;
for (n = 0; n < gp->nr_cpus; n++) {
uint64_t track[2];
fd[n] = perf_event_open(&attr, -1, n, -1, 0);
if (fd[n] == -1)
return errno;
/* read back the event to establish id->tracepoint */
if (read(fd[n], track, sizeof(track)) < 0)
return errno;
sample[n].id = track[1];
sample[n].func = func;
}
gp->nr_events++;
return 0;
}
static int perf_mmap(struct gpu_perf *gp)
{
int size = (1 + N_PAGES) * gp->page_size;
int *fd, i, j;
gp->map = malloc(sizeof(void *)*gp->nr_cpus);
if (gp->map == NULL)
return ENOMEM;
fd = gp->fd;
for (j = 0; j < gp->nr_cpus; j++) {
gp->map[j] = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_SHARED, *fd++, 0);
if (gp->map[j] == (void *)-1)
goto err;
}
for (i = 1; i < gp->nr_events; i++) {
for (j = 0; j < gp->nr_cpus; j++)
ioctl(*fd++, PERF_EVENT_IOC_SET_OUTPUT, gp->fd[j]);
}
return 0;
err:
while (--j > 0)
munmap(gp->map[j], size);
free(gp->map);
gp->map = NULL;
return EINVAL;
}
static char *get_comm(pid_t pid, char *comm, int len)
{
char filename[1024];
int fd;
*comm = '\0';
snprintf(filename, sizeof(filename), "/proc/%d/comm", pid);
fd = open(filename, 0);
if (fd >= 0) {
len = read(fd, comm, len-1);
if (len >= 0)
comm[len-1] = '\0';
close(fd);
}
return comm;
}
static struct gpu_perf_comm *
lookup_comm(struct gpu_perf *gp, pid_t pid)
{
struct gpu_perf_comm *comm;
for (comm = gp->comm; comm != NULL; comm = comm->next) {
if (comm->pid == pid)
break;
}
if (comm == NULL) {
comm = calloc(1, sizeof(*comm));
if (comm == NULL)
return NULL;
comm->next = gp->comm;
gp->comm = comm;
get_comm(pid, comm->name, sizeof(comm->name));
comm->pid = pid;
}
return comm;
}
static int request_add(struct gpu_perf *gp, const void *event)
{
const struct sample_event *sample = event;
struct gpu_perf_comm *comm;
comm = lookup_comm(gp, sample->pid);
if (comm == NULL)
return 0;
comm->nr_requests[sample->raw[1]]++;
return 1;
}
static int seqno_start(struct gpu_perf *gp, const void *event)
{
return 0;
}
static int seqno_end(struct gpu_perf *gp, const void *event)
{
return 0;
}
static int flip_complete(struct gpu_perf *gp, const void *event)
{
gp->flip_complete++;
return 1;
}
static int wait_begin(struct gpu_perf *gp, const void *event)
{
const struct sample_event *sample = event;
struct gpu_perf_comm *comm;
struct gpu_perf_wait *wait;
comm = lookup_comm(gp, sample->pid);
if (comm == NULL)
return 0;
wait = malloc(sizeof(*wait));
if (wait == NULL)
return 0;
wait->seqno = sample->raw[3];
wait->time = sample->time;
wait->next = comm->wait;
comm->wait = wait;
return 0;
}
static int wait_end(struct gpu_perf *gp, const void *event)
{
const struct sample_event *sample = event;
struct gpu_perf_comm *comm;
struct gpu_perf_wait *wait, **prev;
comm = lookup_comm(gp, sample->pid);
if (comm == NULL)
return 0;
for (prev = &comm->wait; (wait = *prev) != NULL; prev = &wait->next) {
if (wait->seqno != sample->raw[3])
continue;
comm->wait_time += sample->time - wait->time;
*prev = wait->next;
free(wait);
return 1;
}
return 0;
}
void gpu_perf_init(struct gpu_perf *gp, unsigned flags)
{
memset(gp, 0, sizeof(*gp));
gp->nr_cpus = sysconf(_SC_NPROCESSORS_ONLN);
gp->page_size = getpagesize();
perf_tracepoint_open(gp, "i915", "i915_gem_request_add", request_add);
if (perf_tracepoint_open(gp, "i915", "i915_gem_ring_complete", seqno_end) == 0)
perf_tracepoint_open(gp, "i915", "i915_gem_ring_dispatch", seqno_start);
if (perf_tracepoint_open(gp, "i915", "i915_gem_request_wait_begin", wait_begin) == 0)
perf_tracepoint_open(gp, "i915", "i915_gem_request_wait_end", wait_end);
perf_tracepoint_open(gp, "i915", "i915_flip_complete", flip_complete);
if (gp->nr_events == 0)
return;
if (perf_mmap(gp))
return;
}
static int process_sample(struct gpu_perf *gp,
const struct perf_event_header *header)
{
const struct sample_event *sample = (const struct sample_event *)header;
int n, update = 0;
/* hash me! */
for (n = 0; n < gp->nr_cpus * gp->nr_events; n++) {
if (gp->sample[n].id != sample->id)
continue;
update = 1;
if (gp->sample[n].func)
update = gp->sample[n].func(gp, sample);
break;
}
return update;
}
int gpu_perf_update(struct gpu_perf *gp)
{
const int size = N_PAGES * gp->page_size;
const int mask = size - 1;
uint8_t *buffer = NULL;
int buffer_size = 0;
int n, update = 0;
if (gp->map == NULL)
return 0;
for (n = 0; n < gp->nr_cpus; n++) {
struct perf_event_mmap_page *mmap = gp->map[n];
const uint8_t *data;
uint64_t head, tail;
int wrap = 0;
tail = mmap->data_tail;
head = mmap->data_head;
rmb();
if (head < tail) {
wrap = 1;
tail &= mask;
head &= mask;
head += size;
}
data = (uint8_t *)mmap + gp->page_size;
while (head - tail >= sizeof (struct perf_event_header)) {
const struct perf_event_header *header;
header = (const struct perf_event_header *)(data + (tail & mask));
if (header->size > head - tail)
break;
if ((const uint8_t *)header + header->size > data + size) {
int before;
if (header->size > buffer_size) {
uint8_t *b = realloc(buffer, header->size);
if (b == NULL)
break;
buffer = b;
buffer_size = header->size;
}
before = data + size - (const uint8_t *)header;
memcpy(buffer, header, before);
memcpy(buffer + before, data, header->size - before);
header = (struct perf_event_header *)buffer;
}
if (header->type == PERF_RECORD_SAMPLE)
update += process_sample(gp, header);
tail += header->size;
}
if (wrap)
tail &= mask;
mmap->data_tail = tail;
}
free(buffer);
return update;
}