igt/gem_bad_reloc: Fix handling of 64bit relocations

Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=94537
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
This commit is contained in:
Chris Wilson 2016-03-14 20:40:53 +00:00
parent f0d5874bd8
commit 1430eb065f

View File

@ -43,106 +43,95 @@ IGT_TEST_DESCRIPTION("Simulates SNA behaviour using negative self-relocations"
" for STATE_BASE_ADDRESS command packets."); " for STATE_BASE_ADDRESS command packets.");
#define USE_LUT (1 << 12) #define USE_LUT (1 << 12)
#define BIAS (256*1024)
/* Simulates SNA behaviour using negative self-relocations for /* Simulates SNA behaviour using negative self-relocations for
* STATE_BASE_ADDRESS command packets. If they wrap around (to values greater * STATE_BASE_ADDRESS command packets. If they wrap around (to values greater
* than the total size of the GTT), the GPU will hang. * than the total size of the GTT), the GPU will hang.
* See https://bugs.freedesktop.org/show_bug.cgi?id=78533 * See https://bugs.freedesktop.org/show_bug.cgi?id=78533
*/ */
static int negative_reloc(int fd, unsigned flags) static void negative_reloc(int fd, unsigned engine, unsigned flags)
{ {
struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_execbuffer2 execbuf;
struct drm_i915_gem_exec_object2 gem_exec[2]; struct drm_i915_gem_exec_object2 obj;
struct drm_i915_gem_relocation_entry gem_reloc[1000]; struct drm_i915_gem_relocation_entry reloc[1000];
uint64_t gtt_max = gem_aperture_size(fd); uint64_t gtt_max = gem_aperture_size(fd);
uint32_t buf[1024] = {MI_BATCH_BUFFER_END}; uint32_t bbe = MI_BATCH_BUFFER_END;
uint64_t *offsets;
int i; int i;
#define BIAS (256*1024) gem_require_ring(fd, engine);
igt_require(intel_gen(intel_get_drm_devid(fd)) >= 7); igt_require(intel_gen(intel_get_drm_devid(fd)) >= 7);
memset(gem_exec, 0, sizeof(gem_exec)); memset(&obj, 0, sizeof(obj));
gem_exec[0].handle = gem_create(fd, 4096); obj.handle = gem_create(fd, 8192);
gem_write(fd, gem_exec[0].handle, 0, buf, 8); gem_write(fd, obj.handle, 0, &bbe, sizeof(bbe));
gem_reloc[0].offset = 1024;
gem_reloc[0].delta = 0;
gem_reloc[0].target_handle = gem_exec[0].handle;
gem_reloc[0].read_domains = I915_GEM_DOMAIN_COMMAND;
gem_exec[1].handle = gem_create(fd, 4096);
gem_write(fd, gem_exec[1].handle, 0, buf, 8);
gem_exec[1].relocation_count = 1;
gem_exec[1].relocs_ptr = (uintptr_t)gem_reloc;
memset(&execbuf, 0, sizeof(execbuf)); memset(&execbuf, 0, sizeof(execbuf));
execbuf.buffers_ptr = (uintptr_t)gem_exec; execbuf.buffers_ptr = (uintptr_t)&obj;
execbuf.buffer_count = 2; execbuf.buffer_count = 1;
execbuf.batch_len = 8; execbuf.flags = engine | (flags & USE_LUT);
igt_require(__gem_execbuf(fd, &execbuf) == 0);
do_ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, &execbuf); igt_info("Found offset %lld for 4k batch\n", (long long)obj.offset);
gem_close(fd, gem_exec[1].handle);
igt_info("Found offset %lld for 4k batch\n", (long long)gem_exec[0].offset);
/* /*
* Ideally we'd like to be able to control where the kernel is going to * Ideally we'd like to be able to control where the kernel is going to
* place the buffer. We don't SKIP here because it causes the test * place the buffer. We don't SKIP here because it causes the test
* to "randomly" flip-flop between the SKIP and PASS states. * to "randomly" flip-flop between the SKIP and PASS states.
*/ */
if (gem_exec[0].offset < BIAS) { if (obj.offset < BIAS) {
igt_info("Offset is below BIAS, not testing anything\n"); igt_info("Offset is below BIAS, not testing anything\n");
return 0; return;
} }
memset(gem_reloc, 0, sizeof(gem_reloc)); memset(reloc, 0, sizeof(reloc));
for (i = 0; i < sizeof(gem_reloc)/sizeof(gem_reloc[0]); i++) { for (i = 0; i < ARRAY_SIZE(reloc); i++) {
gem_reloc[i].offset = 8 + 4*i; reloc[i].offset = 8 + 8*i;
gem_reloc[i].delta = -BIAS*i/1024; reloc[i].delta = -BIAS*i/1024;
gem_reloc[i].target_handle = flags & USE_LUT ? 0 : gem_exec[0].handle; reloc[i].presumed_offset = -1;
gem_reloc[i].read_domains = I915_GEM_DOMAIN_COMMAND; reloc[i].target_handle = flags & USE_LUT ? 0 : obj.handle;
reloc[i].read_domains = I915_GEM_DOMAIN_COMMAND;
} }
obj.relocation_count = i;
obj.relocs_ptr = (uintptr_t)reloc;
gem_execbuf(fd, &execbuf);
gem_exec[0].relocation_count = sizeof(gem_reloc)/sizeof(gem_reloc[0]); igt_info("Batch is now at offset %#llx, max GTT %#llx\n",
gem_exec[0].relocs_ptr = (uintptr_t)gem_reloc; (long long)obj.offset, (long long)gtt_max);
execbuf.buffer_count = 1; offsets = gem_mmap__cpu(fd, obj.handle, 0, 8192, PROT_READ);
execbuf.flags = flags & USE_LUT; gem_set_domain(fd, obj.handle, I915_GEM_DOMAIN_CPU, 0);
do_ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, &execbuf); gem_close(fd, obj.handle);
igt_info("Batch is now at offset %lld\n", (long long)gem_exec[0].offset); for (i = 0; i < ARRAY_SIZE(reloc); i++)
igt_assert_f(offsets[1 + i] < gtt_max,
gem_read(fd, gem_exec[0].handle, 0, buf, sizeof(buf)); "Offset[%d]=%#llx, expected less than %#llx\n",
gem_close(fd, gem_exec[0].handle); i, (long long)offsets[i+i], (long long)gtt_max);
munmap(offsets, 8192);
for (i = 0; i < sizeof(gem_reloc)/sizeof(gem_reloc[0]); i++)
igt_assert(buf[2 + i] < gtt_max);
return 0;
} }
static int negative_reloc_blt(int fd) static void negative_reloc_blt(int fd)
{ {
const int gen = intel_gen(intel_get_drm_devid(fd)); const int gen = intel_gen(intel_get_drm_devid(fd));
struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_execbuffer2 execbuf;
struct drm_i915_gem_exec_object2 gem_exec[1024][2]; struct drm_i915_gem_exec_object2 obj[1024][2];
struct drm_i915_gem_relocation_entry gem_reloc; struct drm_i915_gem_relocation_entry reloc;
uint32_t buf[1024], *b; uint32_t buf[1024], *b;
int i; int i;
memset(&gem_reloc, 0, sizeof(gem_reloc)); memset(&reloc, 0, sizeof(reloc));
gem_reloc.offset = 4 * sizeof(uint32_t); reloc.offset = 4 * sizeof(uint32_t);
gem_reloc.presumed_offset = ~0ULL; reloc.presumed_offset = ~0ULL;
gem_reloc.delta = -4096; reloc.delta = -4096;
gem_reloc.target_handle = 0; reloc.target_handle = 0;
gem_reloc.read_domains = I915_GEM_DOMAIN_RENDER; reloc.read_domains = I915_GEM_DOMAIN_RENDER;
gem_reloc.write_domain = I915_GEM_DOMAIN_RENDER; reloc.write_domain = I915_GEM_DOMAIN_RENDER;
for (i = 0; i < 1024; i++) { for (i = 0; i < 1024; i++) {
memset(gem_exec[i], 0, sizeof(gem_exec[i])); memset(obj[i], 0, sizeof(obj[i]));
gem_exec[i][0].handle = gem_create(fd, 4096); obj[i][0].handle = gem_create(fd, 4096);
gem_exec[i][0].flags = EXEC_OBJECT_NEEDS_FENCE; obj[i][0].flags = EXEC_OBJECT_NEEDS_FENCE;
b = buf; b = buf;
*b++ = XY_COLOR_BLT_CMD_NOLEN | *b++ = XY_COLOR_BLT_CMD_NOLEN |
@ -159,10 +148,10 @@ static int negative_reloc_blt(int fd)
if ((b - buf) & 1) if ((b - buf) & 1)
*b++ = 0; *b++ = 0;
gem_exec[i][1].handle = gem_create(fd, 4096); obj[i][1].handle = gem_create(fd, 4096);
gem_write(fd, gem_exec[i][1].handle, 0, buf, (b - buf) * sizeof(uint32_t)); gem_write(fd, obj[i][1].handle, 0, buf, (b - buf) * sizeof(uint32_t));
gem_exec[i][1].relocation_count = 1; obj[i][1].relocation_count = 1;
gem_exec[i][1].relocs_ptr = (uintptr_t)&gem_reloc; obj[i][1].relocs_ptr = (uintptr_t)&reloc;
} }
memset(&execbuf, 0, sizeof(execbuf)); memset(&execbuf, 0, sizeof(execbuf));
@ -173,15 +162,15 @@ static int negative_reloc_blt(int fd)
execbuf.flags |= I915_EXEC_BLT; execbuf.flags |= I915_EXEC_BLT;
for (i = 0; i < 1024; i++) { for (i = 0; i < 1024; i++) {
execbuf.buffers_ptr = (uintptr_t)gem_exec[i]; execbuf.buffers_ptr = (uintptr_t)obj[i];
gem_execbuf(fd, &execbuf); gem_execbuf(fd, &execbuf);
} }
for (i = 1024; i--;) { for (i = 1024; i--;) {
gem_read(fd, gem_exec[i][0].handle, gem_read(fd, obj[i][0].handle,
i*sizeof(uint32_t), buf + i, sizeof(uint32_t)); i*sizeof(uint32_t), buf + i, sizeof(uint32_t));
gem_close(fd, gem_exec[i][0].handle); gem_close(fd, obj[i][0].handle);
gem_close(fd, gem_exec[i][1].handle); gem_close(fd, obj[i][1].handle);
} }
if (0) { if (0) {
@ -192,28 +181,27 @@ static int negative_reloc_blt(int fd)
} }
for (i = 0; i < 1024; i++) for (i = 0; i < 1024; i++)
igt_assert_eq(buf[i], 0xc0ffee ^ i); igt_assert_eq(buf[i], 0xc0ffee ^ i);
return 0;
} }
int fd;
igt_main igt_main
{ {
igt_fixture { const struct intel_execution_engine *e;
int fd = -1;
igt_fixture
fd = drm_open_driver(DRIVER_INTEL); fd = drm_open_driver(DRIVER_INTEL);
for (e = intel_execution_engines; e->name; e++) {
igt_subtest_f("negative-reloc-%s", e->name)
negative_reloc(fd, e->exec_id | e->flags, 0);
igt_subtest_f("negative-reloc-lut-%s", e->name)
negative_reloc(fd, e->exec_id | e->flags, USE_LUT);
} }
igt_subtest("negative-reloc")
negative_reloc(fd, 0);
igt_subtest("negative-reloc-lut")
negative_reloc(fd, USE_LUT);
igt_subtest("negative-reloc-blt") igt_subtest("negative-reloc-blt")
negative_reloc_blt(fd); negative_reloc_blt(fd);
igt_fixture { igt_fixture
close(fd); close(fd);
}
} }