mirror of
https://github.com/tiagovignatti/intel-gpu-tools.git
synced 2025-06-08 08:26:10 +00:00
Add a header that includes all the headers for the library. This allows reorganisation of the library without affecting programs using it and also simplifies the headers that need to be included to use the library. Signed-off-by: Thomas Wood <thomas.wood@intel.com>
486 lines
10 KiB
C
486 lines
10 KiB
C
/*
|
|
* Copyright © 2011 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Chris Wilson <chris@chris-wilson.co.uk>
|
|
*
|
|
*/
|
|
|
|
#define _GNU_SOURCE
|
|
#include "igt.h"
|
|
#include <unistd.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <fcntl.h>
|
|
#include <inttypes.h>
|
|
#include <pthread.h>
|
|
#include <errno.h>
|
|
#include <sys/stat.h>
|
|
#include <sys/ioctl.h>
|
|
#include "drm.h"
|
|
|
|
struct local_i915_gem_mmap_v2 {
|
|
uint32_t handle;
|
|
uint32_t pad;
|
|
uint64_t offset;
|
|
uint64_t size;
|
|
uint64_t addr_ptr;
|
|
uint64_t flags;
|
|
#define I915_MMAP_WC 0x1
|
|
};
|
|
#define LOCAL_IOCTL_I915_GEM_MMAP_v2 DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_MMAP, struct local_i915_gem_mmap_v2)
|
|
|
|
static int OBJECT_SIZE = 16*1024*1024;
|
|
|
|
static void set_domain(int fd, uint32_t handle)
|
|
{
|
|
gem_set_domain(fd, handle, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
|
|
}
|
|
|
|
static void *
|
|
mmap_bo(int fd, uint32_t handle)
|
|
{
|
|
void *ptr;
|
|
|
|
ptr = gem_mmap__wc(fd, handle, 0, OBJECT_SIZE, PROT_READ | PROT_WRITE);
|
|
igt_assert(ptr && ptr != MAP_FAILED);
|
|
|
|
return ptr;
|
|
}
|
|
|
|
static void *
|
|
create_pointer(int fd)
|
|
{
|
|
uint32_t handle;
|
|
void *ptr;
|
|
|
|
handle = gem_create(fd, OBJECT_SIZE);
|
|
|
|
ptr = mmap_bo(fd, handle);
|
|
set_domain(fd, handle);
|
|
|
|
gem_close(fd, handle);
|
|
|
|
return ptr;
|
|
}
|
|
|
|
static void
|
|
test_invalid_flags(int fd)
|
|
{
|
|
struct drm_i915_getparam gp;
|
|
struct local_i915_gem_mmap_v2 arg;
|
|
uint64_t flag = I915_MMAP_WC;
|
|
int val = -1;
|
|
|
|
memset(&arg, 0, sizeof(arg));
|
|
arg.handle = gem_create(fd, 4096);
|
|
arg.offset = 0;
|
|
arg.size = 4096;
|
|
|
|
memset(&gp, 0, sizeof(gp));
|
|
gp.param = 30; /* MMAP_VERSION */
|
|
gp.value = &val;
|
|
|
|
/* Do we have the new mmap_ioctl? */
|
|
drmIoctl(fd, DRM_IOCTL_I915_GETPARAM, &gp);
|
|
|
|
if (val >= 1) {
|
|
/*
|
|
* Only MMAP_WC flag is supported in version 1, so any other
|
|
* flag should be rejected.
|
|
*/
|
|
flag <<= 1;
|
|
while (flag) {
|
|
arg.flags = flag;
|
|
igt_assert(drmIoctl(fd,
|
|
LOCAL_IOCTL_I915_GEM_MMAP_v2,
|
|
&arg) == -1);
|
|
igt_assert_eq(errno, EINVAL);
|
|
flag <<= 1;
|
|
}
|
|
}
|
|
|
|
gem_close(fd, arg.handle);
|
|
}
|
|
|
|
static void
|
|
test_copy(int fd)
|
|
{
|
|
void *src, *dst;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
/* copy from a fresh src to fresh dst to force pagefault on both */
|
|
src = create_pointer(fd);
|
|
dst = create_pointer(fd);
|
|
|
|
memcpy(dst, src, OBJECT_SIZE);
|
|
memcpy(src, dst, OBJECT_SIZE);
|
|
|
|
munmap(dst, OBJECT_SIZE);
|
|
munmap(src, OBJECT_SIZE);
|
|
}
|
|
|
|
enum test_read_write {
|
|
READ_BEFORE_WRITE,
|
|
READ_AFTER_WRITE,
|
|
};
|
|
|
|
static void
|
|
test_read_write(int fd, enum test_read_write order)
|
|
{
|
|
uint32_t handle;
|
|
void *ptr;
|
|
volatile uint32_t val = 0;
|
|
|
|
handle = gem_create(fd, OBJECT_SIZE);
|
|
set_domain(fd, handle);
|
|
|
|
ptr = mmap_bo(fd, handle);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
|
|
if (order == READ_BEFORE_WRITE) {
|
|
val = *(uint32_t *)ptr;
|
|
*(uint32_t *)ptr = val;
|
|
} else {
|
|
*(uint32_t *)ptr = val;
|
|
val = *(uint32_t *)ptr;
|
|
}
|
|
|
|
gem_close(fd, handle);
|
|
munmap(ptr, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_read_write2(int fd, enum test_read_write order)
|
|
{
|
|
uint32_t handle;
|
|
void *r, *w;
|
|
volatile uint32_t val = 0;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
handle = gem_create(fd, OBJECT_SIZE);
|
|
set_domain(fd, handle);
|
|
|
|
r = gem_mmap__wc(fd, handle, 0, OBJECT_SIZE, PROT_READ);
|
|
igt_assert(r != MAP_FAILED);
|
|
|
|
w = gem_mmap__wc(fd, handle, 0, OBJECT_SIZE, PROT_READ | PROT_WRITE);
|
|
igt_assert(w != MAP_FAILED);
|
|
|
|
if (order == READ_BEFORE_WRITE) {
|
|
val = *(uint32_t *)r;
|
|
*(uint32_t *)w = val;
|
|
} else {
|
|
*(uint32_t *)w = val;
|
|
val = *(uint32_t *)r;
|
|
}
|
|
|
|
gem_close(fd, handle);
|
|
munmap(r, OBJECT_SIZE);
|
|
munmap(w, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_write(int fd)
|
|
{
|
|
void *src;
|
|
uint32_t dst;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
/* copy from a fresh src to fresh dst to force pagefault on both */
|
|
src = create_pointer(fd);
|
|
dst = gem_create(fd, OBJECT_SIZE);
|
|
|
|
gem_write(fd, dst, 0, src, OBJECT_SIZE);
|
|
|
|
gem_close(fd, dst);
|
|
munmap(src, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_write_gtt(int fd)
|
|
{
|
|
uint32_t dst;
|
|
char *dst_gtt;
|
|
void *src;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
dst = gem_create(fd, OBJECT_SIZE);
|
|
set_domain(fd, dst);
|
|
|
|
/* prefault object into gtt */
|
|
dst_gtt = mmap_bo(fd, dst);
|
|
memset(dst_gtt, 0, OBJECT_SIZE);
|
|
munmap(dst_gtt, OBJECT_SIZE);
|
|
|
|
src = create_pointer(fd);
|
|
|
|
gem_write(fd, dst, 0, src, OBJECT_SIZE);
|
|
|
|
gem_close(fd, dst);
|
|
munmap(src, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_read(int fd)
|
|
{
|
|
void *dst;
|
|
uint32_t src;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
/* copy from a fresh src to fresh dst to force pagefault on both */
|
|
dst = create_pointer(fd);
|
|
src = gem_create(fd, OBJECT_SIZE);
|
|
|
|
gem_read(fd, src, 0, dst, OBJECT_SIZE);
|
|
|
|
gem_close(fd, src);
|
|
munmap(dst, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_close(int fd)
|
|
{
|
|
uint32_t handle = gem_create(fd, OBJECT_SIZE);
|
|
uint8_t *ptr = mmap_bo(fd, handle);
|
|
int i;
|
|
|
|
memset(ptr, 0xcc, OBJECT_SIZE);
|
|
gem_close(fd, handle);
|
|
for (i = 0; i < 4096; i++)
|
|
igt_assert(ptr[i*4096+i] == 0xcc);
|
|
|
|
munmap(ptr, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_write_cpu_read_wc(int fd, int force_domain)
|
|
{
|
|
uint32_t handle;
|
|
uint32_t *src, *dst;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
handle = gem_create(fd, OBJECT_SIZE);
|
|
|
|
dst = gem_mmap__wc(fd, handle, 0, OBJECT_SIZE, PROT_READ);
|
|
igt_assert(dst != (uint32_t *)MAP_FAILED);
|
|
|
|
src = gem_mmap__cpu(fd, handle, 0, OBJECT_SIZE, PROT_WRITE);
|
|
igt_assert(src != (uint32_t *)MAP_FAILED);
|
|
|
|
memset(src, 0xaa, OBJECT_SIZE);
|
|
if (force_domain)
|
|
set_domain(fd, handle);
|
|
igt_assert(memcmp(dst, src, OBJECT_SIZE) == 0);
|
|
gem_close(fd, handle);
|
|
|
|
munmap(src, OBJECT_SIZE);
|
|
munmap(dst, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_write_gtt_read_wc(int fd)
|
|
{
|
|
uint32_t handle;
|
|
uint32_t *src, *dst;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
handle = gem_create(fd, OBJECT_SIZE);
|
|
set_domain(fd, handle);
|
|
|
|
dst = gem_mmap__wc(fd, handle, 0, OBJECT_SIZE, PROT_READ);
|
|
igt_assert(dst != (uint32_t *)MAP_FAILED);
|
|
|
|
src = gem_mmap__gtt(fd, handle, OBJECT_SIZE, PROT_WRITE);
|
|
igt_assert(src != (uint32_t *)MAP_FAILED);
|
|
|
|
memset(src, 0xaa, OBJECT_SIZE);
|
|
igt_assert(memcmp(dst, src, OBJECT_SIZE) == 0);
|
|
gem_close(fd, handle);
|
|
|
|
munmap(src, OBJECT_SIZE);
|
|
munmap(dst, OBJECT_SIZE);
|
|
}
|
|
|
|
static void
|
|
test_set_cache_level(int fd)
|
|
{
|
|
struct drm_mode_cursor arg;
|
|
struct drm_mode_card_res res;
|
|
uint32_t crtc[32];
|
|
int active_crtc = 0;
|
|
int n;
|
|
|
|
/* We want to trigger an old WARN in set-cache-level when
|
|
* it sees an unbound object in the GTT domain, following
|
|
* the introduction of mmap(wc).
|
|
*/
|
|
|
|
memset(&arg, 0, sizeof(arg));
|
|
arg.flags = DRM_MODE_CURSOR_BO;
|
|
arg.width = arg.height = 64;
|
|
arg.handle = gem_create(fd, 64*64*4);
|
|
set_domain(fd, arg.handle);
|
|
|
|
/* Bind the object to the cursor to force set-cache-level(DISPLAY) */
|
|
memset(&res, 0, sizeof(res));
|
|
res.count_crtcs = 32;
|
|
res.crtc_id_ptr = (uintptr_t)crtc;
|
|
do_ioctl(fd, DRM_IOCTL_MODE_GETRESOURCES, &res);
|
|
for (n = 0; n < res.count_crtcs; n++) {
|
|
struct drm_mode_crtc mode;
|
|
|
|
memset(&mode, 0, sizeof(mode));
|
|
mode.crtc_id = crtc[n];
|
|
do_ioctl(fd, DRM_IOCTL_MODE_GETCRTC, &mode);
|
|
|
|
if (!mode.mode_valid)
|
|
continue;
|
|
|
|
active_crtc++;
|
|
|
|
arg.crtc_id = crtc[n];
|
|
do_ioctl(fd, DRM_IOCTL_MODE_CURSOR, &arg);
|
|
}
|
|
|
|
gem_close(fd, arg.handle);
|
|
igt_require(active_crtc);
|
|
}
|
|
|
|
struct thread_fault_concurrent {
|
|
pthread_t thread;
|
|
int id;
|
|
uint32_t **ptr;
|
|
};
|
|
|
|
static void *
|
|
thread_fault_concurrent(void *closure)
|
|
{
|
|
struct thread_fault_concurrent *t = closure;
|
|
uint32_t val = 0;
|
|
int n;
|
|
|
|
for (n = 0; n < 32; n++) {
|
|
if (n & 1)
|
|
*t->ptr[(n + t->id) % 32] = val;
|
|
else
|
|
val = *t->ptr[(n + t->id) % 32];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
test_fault_concurrent(int fd)
|
|
{
|
|
uint32_t *ptr[32];
|
|
struct thread_fault_concurrent thread[64];
|
|
int n;
|
|
|
|
gem_require_mmap_wc(fd);
|
|
|
|
for (n = 0; n < 32; n++) {
|
|
ptr[n] = create_pointer(fd);
|
|
}
|
|
|
|
for (n = 0; n < 64; n++) {
|
|
thread[n].ptr = ptr;
|
|
thread[n].id = n;
|
|
pthread_create(&thread[n].thread, NULL, thread_fault_concurrent, &thread[n]);
|
|
}
|
|
|
|
for (n = 0; n < 64; n++)
|
|
pthread_join(thread[n].thread, NULL);
|
|
|
|
for (n = 0; n < 32; n++) {
|
|
munmap(ptr[n], OBJECT_SIZE);
|
|
}
|
|
}
|
|
|
|
static void
|
|
run_without_prefault(int fd,
|
|
void (*func)(int fd))
|
|
{
|
|
igt_disable_prefault();
|
|
func(fd);
|
|
igt_enable_prefault();
|
|
}
|
|
|
|
int fd;
|
|
|
|
igt_main
|
|
{
|
|
if (igt_run_in_simulation())
|
|
OBJECT_SIZE = 1 * 1024 * 1024;
|
|
|
|
igt_fixture
|
|
fd = drm_open_any();
|
|
|
|
igt_subtest("invalid-flags")
|
|
test_invalid_flags(fd);
|
|
igt_subtest("close")
|
|
test_close(fd);
|
|
igt_subtest("copy")
|
|
test_copy(fd);
|
|
igt_subtest("read")
|
|
test_read(fd);
|
|
igt_subtest("write")
|
|
test_write(fd);
|
|
igt_subtest("write-gtt")
|
|
test_write_gtt(fd);
|
|
igt_subtest("read-write")
|
|
test_read_write(fd, READ_BEFORE_WRITE);
|
|
igt_subtest("write-read")
|
|
test_read_write(fd, READ_AFTER_WRITE);
|
|
igt_subtest("read-write-distinct")
|
|
test_read_write2(fd, READ_BEFORE_WRITE);
|
|
igt_subtest("write-read-distinct")
|
|
test_read_write2(fd, READ_AFTER_WRITE);
|
|
igt_subtest("fault-concurrent")
|
|
test_fault_concurrent(fd);
|
|
igt_subtest("read-no-prefault")
|
|
run_without_prefault(fd, test_read);
|
|
igt_subtest("write-no-prefault")
|
|
run_without_prefault(fd, test_write);
|
|
igt_subtest("write-gtt-no-prefault")
|
|
run_without_prefault(fd, test_write_gtt);
|
|
igt_subtest("write-cpu-read-wc")
|
|
test_write_cpu_read_wc(fd, 1);
|
|
igt_subtest("write-cpu-read-wc-unflushed")
|
|
test_write_cpu_read_wc(fd, 0);
|
|
igt_subtest("write-gtt-read-wc")
|
|
test_write_gtt_read_wc(fd);
|
|
igt_subtest("set-cache-level")
|
|
test_set_cache_level(fd);
|
|
|
|
igt_fixture
|
|
close(fd);
|
|
}
|