mirror of
				https://github.com/tiagovignatti/intel-gpu-tools.git
				synced 2025-11-04 12:07:12 +00:00 
			
		
		
		
	Since we allocate 2 times count number of objects, that is the number of objects we should be using when checking for sufficient RAM to use for the test. References: https://bugs.freedesktop.org/show_bug.cgi?id=92799 Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
		
			
				
	
	
		
			1554 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1554 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright © 2009-2014 Intel Corporation
 | 
						|
 *
 | 
						|
 * Permission is hereby granted, free of charge, to any person obtaining a
 | 
						|
 * copy of this software and associated documentation files (the "Software"),
 | 
						|
 * to deal in the Software without restriction, including without limitation
 | 
						|
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 | 
						|
 * and/or sell copies of the Software, and to permit persons to whom the
 | 
						|
 * Software is furnished to do so, subject to the following conditions:
 | 
						|
 *
 | 
						|
 * The above copyright notice and this permission notice (including the next
 | 
						|
 * paragraph) shall be included in all copies or substantial portions of the
 | 
						|
 * Software.
 | 
						|
 *
 | 
						|
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | 
						|
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | 
						|
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 | 
						|
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | 
						|
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 | 
						|
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 | 
						|
 * IN THE SOFTWARE.
 | 
						|
 *
 | 
						|
 * Authors:
 | 
						|
 *    Eric Anholt <eric@anholt.net>
 | 
						|
 *    Chris Wilson <chris@chris-wilson.co.uk>
 | 
						|
 *    Tvrtko Ursulin <tvrtko.ursulin@intel.com>
 | 
						|
 *
 | 
						|
 */
 | 
						|
 | 
						|
/** @file gem_userptr_blits.c
 | 
						|
 *
 | 
						|
 * This is a test of doing many blits using a mixture of normal system pages
 | 
						|
 * and uncached linear buffers, with a working set larger than the
 | 
						|
 * aperture size.
 | 
						|
 *
 | 
						|
 * The goal is to simply ensure the basics work.
 | 
						|
 */
 | 
						|
 | 
						|
#include "igt.h"
 | 
						|
#include <stdlib.h>
 | 
						|
#include <stdio.h>
 | 
						|
#include <string.h>
 | 
						|
#include <fcntl.h>
 | 
						|
#include <inttypes.h>
 | 
						|
#include <errno.h>
 | 
						|
#include <sys/stat.h>
 | 
						|
#include <sys/time.h>
 | 
						|
#include <sys/mman.h>
 | 
						|
#include <signal.h>
 | 
						|
#include <pthread.h>
 | 
						|
#include <time.h>
 | 
						|
 | 
						|
#include "drm.h"
 | 
						|
#include "i915_drm.h"
 | 
						|
 | 
						|
#include "intel_bufmgr.h"
 | 
						|
 | 
						|
#include "eviction_common.c"
 | 
						|
 | 
						|
#ifndef PAGE_SIZE
 | 
						|
#define PAGE_SIZE 4096
 | 
						|
#endif
 | 
						|
 | 
						|
#define LOCAL_I915_GEM_USERPTR       0x33
 | 
						|
#define LOCAL_IOCTL_I915_GEM_USERPTR DRM_IOWR (DRM_COMMAND_BASE + LOCAL_I915_GEM_USERPTR, struct local_i915_gem_userptr)
 | 
						|
struct local_i915_gem_userptr {
 | 
						|
	uint64_t user_ptr;
 | 
						|
	uint64_t user_size;
 | 
						|
	uint32_t flags;
 | 
						|
#define LOCAL_I915_USERPTR_READ_ONLY (1<<0)
 | 
						|
#define LOCAL_I915_USERPTR_UNSYNCHRONIZED (1<<31)
 | 
						|
	uint32_t handle;
 | 
						|
};
 | 
						|
 | 
						|
static uint32_t userptr_flags = LOCAL_I915_USERPTR_UNSYNCHRONIZED;
 | 
						|
 | 
						|
#define WIDTH 512
 | 
						|
#define HEIGHT 512
 | 
						|
 | 
						|
static uint32_t linear[WIDTH*HEIGHT];
 | 
						|
 | 
						|
static void gem_userptr_test_unsynchronized(void)
 | 
						|
{
 | 
						|
	userptr_flags = LOCAL_I915_USERPTR_UNSYNCHRONIZED;
 | 
						|
}
 | 
						|
 | 
						|
static void gem_userptr_test_synchronized(void)
 | 
						|
{
 | 
						|
	userptr_flags = 0;
 | 
						|
}
 | 
						|
 | 
						|
static int gem_userptr(int fd, void *ptr, int size, int read_only, uint32_t *handle)
 | 
						|
{
 | 
						|
	struct local_i915_gem_userptr userptr;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	memset(&userptr, 0, sizeof(userptr));
 | 
						|
	userptr.user_ptr = (uintptr_t)ptr;
 | 
						|
	userptr.user_size = size;
 | 
						|
	userptr.flags = userptr_flags;
 | 
						|
	if (read_only)
 | 
						|
		userptr.flags |= LOCAL_I915_USERPTR_READ_ONLY;
 | 
						|
 | 
						|
	ret = drmIoctl(fd, LOCAL_IOCTL_I915_GEM_USERPTR, &userptr);
 | 
						|
	if (ret)
 | 
						|
		ret = errno;
 | 
						|
	igt_skip_on_f(ret == ENODEV &&
 | 
						|
		      (userptr_flags & LOCAL_I915_USERPTR_UNSYNCHRONIZED) == 0 &&
 | 
						|
		      !read_only,
 | 
						|
		      "Skipping, synchronized mappings with no kernel CONFIG_MMU_NOTIFIER?");
 | 
						|
	if (ret == 0)
 | 
						|
		*handle = userptr.handle;
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
static void gem_userptr_sync(int fd, uint32_t handle)
 | 
						|
{
 | 
						|
	gem_set_domain(fd, handle, I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
 | 
						|
}
 | 
						|
 | 
						|
static void
 | 
						|
copy(int fd, uint32_t dst, uint32_t src, unsigned int error)
 | 
						|
{
 | 
						|
	uint32_t batch[12];
 | 
						|
	struct drm_i915_gem_relocation_entry reloc[2];
 | 
						|
	struct drm_i915_gem_exec_object2 obj[3];
 | 
						|
	struct drm_i915_gem_execbuffer2 exec;
 | 
						|
	uint32_t handle;
 | 
						|
	int ret, i=0;
 | 
						|
 | 
						|
	batch[i++] = XY_SRC_COPY_BLT_CMD |
 | 
						|
		  XY_SRC_COPY_BLT_WRITE_ALPHA |
 | 
						|
		  XY_SRC_COPY_BLT_WRITE_RGB;
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i - 1] |= 8;
 | 
						|
	else
 | 
						|
		batch[i - 1] |= 6;
 | 
						|
 | 
						|
	batch[i++] = (3 << 24) | /* 32 bits */
 | 
						|
		  (0xcc << 16) | /* copy ROP */
 | 
						|
		  WIDTH*4;
 | 
						|
	batch[i++] = 0; /* dst x1,y1 */
 | 
						|
	batch[i++] = (HEIGHT << 16) | WIDTH; /* dst x2,y2 */
 | 
						|
	batch[i++] = 0; /* dst reloc */
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i++] = 0;
 | 
						|
	batch[i++] = 0; /* src x1,y1 */
 | 
						|
	batch[i++] = WIDTH*4;
 | 
						|
	batch[i++] = 0; /* src reloc */
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i++] = 0;
 | 
						|
	batch[i++] = MI_BATCH_BUFFER_END;
 | 
						|
	batch[i++] = MI_NOOP;
 | 
						|
 | 
						|
	handle = gem_create(fd, 4096);
 | 
						|
	gem_write(fd, handle, 0, batch, sizeof(batch));
 | 
						|
 | 
						|
	reloc[0].target_handle = dst;
 | 
						|
	reloc[0].delta = 0;
 | 
						|
	reloc[0].offset = 4 * sizeof(batch[0]);
 | 
						|
	reloc[0].presumed_offset = 0;
 | 
						|
	reloc[0].read_domains = I915_GEM_DOMAIN_RENDER;
 | 
						|
	reloc[0].write_domain = I915_GEM_DOMAIN_RENDER;
 | 
						|
 | 
						|
	reloc[1].target_handle = src;
 | 
						|
	reloc[1].delta = 0;
 | 
						|
	reloc[1].offset = 7 * sizeof(batch[0]);
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		reloc[1].offset += sizeof(batch[0]);
 | 
						|
	reloc[1].presumed_offset = 0;
 | 
						|
	reloc[1].read_domains = I915_GEM_DOMAIN_RENDER;
 | 
						|
	reloc[1].write_domain = 0;
 | 
						|
 | 
						|
	memset(obj, 0, sizeof(obj));
 | 
						|
	exec.buffer_count = 0;
 | 
						|
	obj[exec.buffer_count++].handle = dst;
 | 
						|
	if (src != dst)
 | 
						|
		obj[exec.buffer_count++].handle = src;
 | 
						|
	obj[exec.buffer_count].handle = handle;
 | 
						|
	obj[exec.buffer_count].relocation_count = 2;
 | 
						|
	obj[exec.buffer_count].relocs_ptr = (uintptr_t)reloc;
 | 
						|
	exec.buffer_count++;
 | 
						|
	exec.buffers_ptr = (uintptr_t)obj;
 | 
						|
 | 
						|
	exec.batch_start_offset = 0;
 | 
						|
	exec.batch_len = i * 4;
 | 
						|
	exec.DR1 = exec.DR4 = 0;
 | 
						|
	exec.num_cliprects = 0;
 | 
						|
	exec.cliprects_ptr = 0;
 | 
						|
	exec.flags = HAS_BLT_RING(intel_get_drm_devid(fd)) ? I915_EXEC_BLT : 0;
 | 
						|
	i915_execbuffer2_set_context_id(exec, 0);
 | 
						|
	exec.rsvd2 = 0;
 | 
						|
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, &exec);
 | 
						|
	if (ret)
 | 
						|
		ret = errno;
 | 
						|
 | 
						|
	if (error == ~0)
 | 
						|
		igt_assert_neq(ret, 0);
 | 
						|
	else
 | 
						|
		igt_assert(ret == error);
 | 
						|
 | 
						|
	gem_close(fd, handle);
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
blit(int fd, uint32_t dst, uint32_t src, uint32_t *all_bo, int n_bo)
 | 
						|
{
 | 
						|
	uint32_t batch[12];
 | 
						|
	struct drm_i915_gem_relocation_entry reloc[2];
 | 
						|
	struct drm_i915_gem_exec_object2 *obj;
 | 
						|
	struct drm_i915_gem_execbuffer2 exec;
 | 
						|
	uint32_t handle;
 | 
						|
	int n, ret, i=0;
 | 
						|
 | 
						|
	batch[i++] = XY_SRC_COPY_BLT_CMD |
 | 
						|
		  XY_SRC_COPY_BLT_WRITE_ALPHA |
 | 
						|
		  XY_SRC_COPY_BLT_WRITE_RGB;
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i - 1] |= 8;
 | 
						|
	else
 | 
						|
		batch[i - 1] |= 6;
 | 
						|
	batch[i++] = (3 << 24) | /* 32 bits */
 | 
						|
		  (0xcc << 16) | /* copy ROP */
 | 
						|
		  WIDTH*4;
 | 
						|
	batch[i++] = 0; /* dst x1,y1 */
 | 
						|
	batch[i++] = (HEIGHT << 16) | WIDTH; /* dst x2,y2 */
 | 
						|
	batch[i++] = 0; /* dst reloc */
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i++] = 0;
 | 
						|
	batch[i++] = 0; /* src x1,y1 */
 | 
						|
	batch[i++] = WIDTH*4;
 | 
						|
	batch[i++] = 0; /* src reloc */
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		batch[i++] = 0;
 | 
						|
	batch[i++] = MI_BATCH_BUFFER_END;
 | 
						|
	batch[i++] = MI_NOOP;
 | 
						|
 | 
						|
	handle = gem_create(fd, 4096);
 | 
						|
	gem_write(fd, handle, 0, batch, sizeof(batch));
 | 
						|
 | 
						|
	reloc[0].target_handle = dst;
 | 
						|
	reloc[0].delta = 0;
 | 
						|
	reloc[0].offset = 4 * sizeof(batch[0]);
 | 
						|
	reloc[0].presumed_offset = 0;
 | 
						|
	reloc[0].read_domains = I915_GEM_DOMAIN_RENDER;
 | 
						|
	reloc[0].write_domain = I915_GEM_DOMAIN_RENDER;
 | 
						|
 | 
						|
	reloc[1].target_handle = src;
 | 
						|
	reloc[1].delta = 0;
 | 
						|
	reloc[1].offset = 7 * sizeof(batch[0]);
 | 
						|
	if (intel_gen(intel_get_drm_devid(fd)) >= 8)
 | 
						|
		reloc[1].offset += sizeof(batch[0]);
 | 
						|
	reloc[1].presumed_offset = 0;
 | 
						|
	reloc[1].read_domains = I915_GEM_DOMAIN_RENDER;
 | 
						|
	reloc[1].write_domain = 0;
 | 
						|
 | 
						|
	obj = calloc(n_bo + 1, sizeof(*obj));
 | 
						|
	for (n = 0; n < n_bo; n++)
 | 
						|
		obj[n].handle = all_bo[n];
 | 
						|
	obj[n].handle = handle;
 | 
						|
	obj[n].relocation_count = 2;
 | 
						|
	obj[n].relocs_ptr = (uintptr_t)reloc;
 | 
						|
 | 
						|
	exec.buffers_ptr = (uintptr_t)obj;
 | 
						|
	exec.buffer_count = n_bo + 1;
 | 
						|
	exec.batch_start_offset = 0;
 | 
						|
	exec.batch_len = i * 4;
 | 
						|
	exec.DR1 = exec.DR4 = 0;
 | 
						|
	exec.num_cliprects = 0;
 | 
						|
	exec.cliprects_ptr = 0;
 | 
						|
	exec.flags = HAS_BLT_RING(intel_get_drm_devid(fd)) ? I915_EXEC_BLT : 0;
 | 
						|
	i915_execbuffer2_set_context_id(exec, 0);
 | 
						|
	exec.rsvd2 = 0;
 | 
						|
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, &exec);
 | 
						|
	if (ret)
 | 
						|
		ret = errno;
 | 
						|
 | 
						|
	gem_close(fd, handle);
 | 
						|
	free(obj);
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static uint32_t
 | 
						|
create_userptr(int fd, uint32_t val, uint32_t *ptr)
 | 
						|
{
 | 
						|
	uint32_t handle;
 | 
						|
	int i, ret;
 | 
						|
 | 
						|
	ret = gem_userptr(fd, ptr, sizeof(linear), 0, &handle);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
	igt_assert(handle != 0);
 | 
						|
 | 
						|
	/* Fill the BO with dwords starting at val */
 | 
						|
	for (i = 0; i < WIDTH*HEIGHT; i++)
 | 
						|
		ptr[i] = val++;
 | 
						|
 | 
						|
	return handle;
 | 
						|
}
 | 
						|
 | 
						|
static void **handle_ptr_map;
 | 
						|
static unsigned *handle_size_map;
 | 
						|
static unsigned int num_handle_map;
 | 
						|
 | 
						|
static void reset_handle_ptr(void)
 | 
						|
{
 | 
						|
	if (num_handle_map == 0)
 | 
						|
		return;
 | 
						|
 | 
						|
	free(handle_ptr_map);
 | 
						|
	handle_ptr_map = NULL;
 | 
						|
 | 
						|
	free(handle_size_map);
 | 
						|
	handle_size_map = NULL;
 | 
						|
 | 
						|
	num_handle_map = 0;
 | 
						|
}
 | 
						|
 | 
						|
static void add_handle_ptr(uint32_t handle, void *ptr, int size)
 | 
						|
{
 | 
						|
	if (handle >= num_handle_map) {
 | 
						|
		int max = (4096 + handle) & -4096;
 | 
						|
 | 
						|
		handle_ptr_map = realloc(handle_ptr_map,
 | 
						|
					 max * sizeof(void*));
 | 
						|
		igt_assert(handle_ptr_map);
 | 
						|
		memset(handle_ptr_map + num_handle_map, 0,
 | 
						|
		       (max - num_handle_map) * sizeof(void*));
 | 
						|
 | 
						|
		handle_size_map = realloc(handle_size_map,
 | 
						|
					  max * sizeof(unsigned));
 | 
						|
		igt_assert(handle_size_map);
 | 
						|
		memset(handle_ptr_map + num_handle_map, 0,
 | 
						|
		       (max - num_handle_map) * sizeof(unsigned));
 | 
						|
 | 
						|
		num_handle_map = max;
 | 
						|
	}
 | 
						|
 | 
						|
	handle_ptr_map[handle] = ptr;
 | 
						|
	handle_size_map[handle] = size;
 | 
						|
}
 | 
						|
 | 
						|
static void *get_handle_ptr(uint32_t handle)
 | 
						|
{
 | 
						|
	igt_assert(handle < num_handle_map);
 | 
						|
	return handle_ptr_map[handle];
 | 
						|
}
 | 
						|
 | 
						|
static void free_handle_ptr(uint32_t handle)
 | 
						|
{
 | 
						|
	igt_assert(handle < num_handle_map);
 | 
						|
	igt_assert(handle_ptr_map[handle]);
 | 
						|
 | 
						|
	munmap(handle_ptr_map[handle], handle_size_map[handle]);
 | 
						|
	handle_ptr_map[handle] = NULL;
 | 
						|
}
 | 
						|
 | 
						|
static uint32_t create_userptr_bo(int fd, uint64_t size)
 | 
						|
{
 | 
						|
	void *ptr;
 | 
						|
	uint32_t handle;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	ptr = mmap(NULL, size,
 | 
						|
		   PROT_READ | PROT_WRITE,
 | 
						|
		   MAP_ANONYMOUS | MAP_SHARED,
 | 
						|
		   -1, 0);
 | 
						|
	igt_assert(ptr != MAP_FAILED);
 | 
						|
 | 
						|
	ret = gem_userptr(fd, (uint32_t *)ptr, size, 0, &handle);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
	add_handle_ptr(handle, ptr, size);
 | 
						|
 | 
						|
	return handle;
 | 
						|
}
 | 
						|
 | 
						|
static void flink_userptr_bo(uint32_t old_handle, uint32_t new_handle)
 | 
						|
{
 | 
						|
	igt_assert(old_handle < num_handle_map);
 | 
						|
	igt_assert(handle_ptr_map[old_handle]);
 | 
						|
 | 
						|
	add_handle_ptr(new_handle,
 | 
						|
		       handle_ptr_map[old_handle],
 | 
						|
		       handle_size_map[old_handle]);
 | 
						|
}
 | 
						|
 | 
						|
static void clear(int fd, uint32_t handle, int size)
 | 
						|
{
 | 
						|
	void *ptr = get_handle_ptr(handle);
 | 
						|
 | 
						|
	igt_assert(ptr != NULL);
 | 
						|
 | 
						|
	memset(ptr, 0, size);
 | 
						|
}
 | 
						|
 | 
						|
static void free_userptr_bo(int fd, uint32_t handle)
 | 
						|
{
 | 
						|
	gem_close(fd, handle);
 | 
						|
	free_handle_ptr(handle);
 | 
						|
}
 | 
						|
 | 
						|
static uint32_t
 | 
						|
create_bo(int fd, uint32_t val)
 | 
						|
{
 | 
						|
	uint32_t handle;
 | 
						|
	int i;
 | 
						|
 | 
						|
	handle = gem_create(fd, sizeof(linear));
 | 
						|
 | 
						|
	/* Fill the BO with dwords starting at val */
 | 
						|
	for (i = 0; i < WIDTH*HEIGHT; i++)
 | 
						|
		linear[i] = val++;
 | 
						|
	gem_write(fd, handle, 0, linear, sizeof(linear));
 | 
						|
 | 
						|
	return handle;
 | 
						|
}
 | 
						|
 | 
						|
static void
 | 
						|
check_cpu(uint32_t *ptr, uint32_t val)
 | 
						|
{
 | 
						|
	int i;
 | 
						|
 | 
						|
	for (i = 0; i < WIDTH*HEIGHT; i++) {
 | 
						|
		igt_assert_f(ptr[i] == val,
 | 
						|
			     "Expected 0x%08x, found 0x%08x "
 | 
						|
			     "at offset 0x%08x\n",
 | 
						|
			     val, ptr[i], i * 4);
 | 
						|
		val++;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static void
 | 
						|
check_gpu(int fd, uint32_t handle, uint32_t val)
 | 
						|
{
 | 
						|
	gem_read(fd, handle, 0, linear, sizeof(linear));
 | 
						|
	check_cpu(linear, val);
 | 
						|
}
 | 
						|
 | 
						|
static int has_userptr(int fd)
 | 
						|
{
 | 
						|
	uint32_t handle = 0;
 | 
						|
	void *ptr;
 | 
						|
	uint32_t oldflags;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
	oldflags = userptr_flags;
 | 
						|
	gem_userptr_test_unsynchronized();
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle);
 | 
						|
	userptr_flags = oldflags;
 | 
						|
	if (ret != 0) {
 | 
						|
		free(ptr);
 | 
						|
		return 0;
 | 
						|
	}
 | 
						|
 | 
						|
	gem_close(fd, handle);
 | 
						|
	free(ptr);
 | 
						|
 | 
						|
	return handle != 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_input_checking(int fd)
 | 
						|
{
 | 
						|
	struct local_i915_gem_userptr userptr;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	/* Invalid flags. */
 | 
						|
	memset(&userptr, 0, sizeof(userptr));
 | 
						|
	userptr.user_ptr = 0;
 | 
						|
	userptr.user_size = 0;
 | 
						|
	userptr.flags = ~0;
 | 
						|
	ret = drmIoctl(fd, LOCAL_IOCTL_I915_GEM_USERPTR, &userptr);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	/* Too big. */
 | 
						|
	memset(&userptr, 0, sizeof(userptr));
 | 
						|
	userptr.user_ptr = 0;
 | 
						|
	userptr.user_size = ~0;
 | 
						|
	userptr.flags = 0;
 | 
						|
	ret = drmIoctl(fd, LOCAL_IOCTL_I915_GEM_USERPTR, &userptr);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	/* Both wrong. */
 | 
						|
	memset(&userptr, 0, sizeof(userptr));
 | 
						|
	userptr.user_ptr = 0;
 | 
						|
	userptr.user_size = ~0;
 | 
						|
	userptr.flags = ~0;
 | 
						|
	ret = drmIoctl(fd, LOCAL_IOCTL_I915_GEM_USERPTR, &userptr);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_access_control(int fd)
 | 
						|
{
 | 
						|
	igt_fork(child, 1) {
 | 
						|
		void *ptr;
 | 
						|
		int ret;
 | 
						|
		uint32_t handle;
 | 
						|
 | 
						|
		igt_drop_root();
 | 
						|
 | 
						|
		/* CAP_SYS_ADMIN is needed for UNSYNCHRONIZED mappings. */
 | 
						|
		gem_userptr_test_unsynchronized();
 | 
						|
 | 
						|
		igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
 | 
						|
		ret = gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle);
 | 
						|
		if (ret == 0)
 | 
						|
			gem_close(fd, handle);
 | 
						|
		free(ptr);
 | 
						|
		igt_assert(ret == EPERM);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_waitchildren();
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_invalid_null_pointer(int fd)
 | 
						|
{
 | 
						|
	uint32_t handle;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	/* NULL pointer. */
 | 
						|
	ret = gem_userptr(fd, NULL, PAGE_SIZE, 0, &handle);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	copy(fd, handle, handle, ~0); /* QQQ Precise errno? */
 | 
						|
	gem_close(fd, handle);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_invalid_gtt_mapping(int fd)
 | 
						|
{
 | 
						|
	uint32_t handle, handle2;
 | 
						|
	void *ptr;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	/* GTT mapping */
 | 
						|
	handle = create_bo(fd, 0);
 | 
						|
	ptr = gem_mmap__gtt(fd, handle, sizeof(linear),
 | 
						|
			    PROT_READ | PROT_WRITE);
 | 
						|
	gem_close(fd, handle);
 | 
						|
	igt_assert(((unsigned long)ptr & (PAGE_SIZE - 1)) == 0);
 | 
						|
	igt_assert((sizeof(linear) & (PAGE_SIZE - 1)) == 0);
 | 
						|
 | 
						|
	ret = gem_userptr(fd, ptr, sizeof(linear), 0, &handle2);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
	copy(fd, handle2, handle2, ~0); /* QQQ Precise errno? */
 | 
						|
	gem_close(fd, handle2);
 | 
						|
 | 
						|
	munmap(ptr, sizeof(linear));
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
#define PE_GTT_MAP 0x1
 | 
						|
#define PE_BUSY 0x2
 | 
						|
static void test_process_exit(int fd, int flags)
 | 
						|
{
 | 
						|
	if (flags & PE_GTT_MAP)
 | 
						|
		igt_require(gem_has_llc(fd));
 | 
						|
 | 
						|
	igt_fork(child, 1) {
 | 
						|
		uint32_t handle;
 | 
						|
 | 
						|
		handle = create_userptr_bo(fd, sizeof(linear));
 | 
						|
 | 
						|
		if (flags & PE_GTT_MAP) {
 | 
						|
			uint32_t *ptr = __gem_mmap__gtt(fd, handle, sizeof(linear), PROT_READ | PROT_WRITE);
 | 
						|
			if (ptr)
 | 
						|
				*ptr = 0;
 | 
						|
		}
 | 
						|
 | 
						|
		if (flags & PE_BUSY)
 | 
						|
			copy(fd, handle, handle, 0);
 | 
						|
	}
 | 
						|
	igt_waitchildren();
 | 
						|
}
 | 
						|
 | 
						|
static void test_forked_access(int fd)
 | 
						|
{
 | 
						|
	uint32_t handle1 = 0, handle2 = 0;
 | 
						|
	void *ptr1 = NULL, *ptr2 = NULL;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	ret = posix_memalign(&ptr1, PAGE_SIZE, sizeof(linear));
 | 
						|
#ifdef MADV_DONTFORK
 | 
						|
	ret |= madvise(ptr1, sizeof(linear), MADV_DONTFORK);
 | 
						|
#endif
 | 
						|
	ret |= gem_userptr(fd, ptr1, sizeof(linear), 0, &handle1);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
	igt_assert(ptr1);
 | 
						|
	igt_assert(handle1);
 | 
						|
 | 
						|
	ret = posix_memalign(&ptr2, PAGE_SIZE, sizeof(linear));
 | 
						|
#ifdef MADV_DONTFORK
 | 
						|
	ret |= madvise(ptr2, sizeof(linear), MADV_DONTFORK);
 | 
						|
#endif
 | 
						|
	ret |= gem_userptr(fd, ptr2, sizeof(linear), 0, &handle2);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
	igt_assert(ptr2);
 | 
						|
	igt_assert(handle2);
 | 
						|
 | 
						|
	memset(ptr1, 0x1, sizeof(linear));
 | 
						|
	memset(ptr2, 0x2, sizeof(linear));
 | 
						|
 | 
						|
	igt_fork(child, 1) {
 | 
						|
		copy(fd, handle1, handle2, 0);
 | 
						|
	}
 | 
						|
	igt_waitchildren();
 | 
						|
 | 
						|
	gem_userptr_sync(fd, handle1);
 | 
						|
	gem_userptr_sync(fd, handle2);
 | 
						|
 | 
						|
	gem_close(fd, handle1);
 | 
						|
	gem_close(fd, handle2);
 | 
						|
 | 
						|
	igt_assert(memcmp(ptr1, ptr2, sizeof(linear)) == 0);
 | 
						|
 | 
						|
#ifdef MADV_DOFORK
 | 
						|
	ret = madvise(ptr1, sizeof(linear), MADV_DOFORK);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
#endif
 | 
						|
	free(ptr1);
 | 
						|
 | 
						|
#ifdef MADV_DOFORK
 | 
						|
	ret = madvise(ptr2, sizeof(linear), MADV_DOFORK);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
#endif
 | 
						|
	free(ptr2);
 | 
						|
}
 | 
						|
 | 
						|
static int test_forbidden_ops(int fd)
 | 
						|
{
 | 
						|
	struct drm_i915_gem_pread gem_pread;
 | 
						|
	struct drm_i915_gem_pwrite gem_pwrite;
 | 
						|
	void *ptr;
 | 
						|
	int ret;
 | 
						|
	uint32_t handle;
 | 
						|
	char buf[PAGE_SIZE];
 | 
						|
 | 
						|
	memset(&gem_pread, 0, sizeof(gem_pread));
 | 
						|
	memset(&gem_pwrite, 0, sizeof(gem_pwrite));
 | 
						|
 | 
						|
	igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	/* pread/pwrite are not always forbidden, but when they
 | 
						|
	 * are they should fail with EINVAL.
 | 
						|
	 */
 | 
						|
 | 
						|
	gem_pread.handle = handle;
 | 
						|
	gem_pread.offset = 0;
 | 
						|
	gem_pread.size = PAGE_SIZE;
 | 
						|
	gem_pread.data_ptr = (uintptr_t)buf;
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_PREAD, &gem_pread);
 | 
						|
	igt_assert(ret == 0 || errno == EINVAL);
 | 
						|
 | 
						|
	gem_pwrite.handle = handle;
 | 
						|
	gem_pwrite.offset = 0;
 | 
						|
	gem_pwrite.size = PAGE_SIZE;
 | 
						|
	gem_pwrite.data_ptr = (uintptr_t)buf;
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_PWRITE, &gem_pwrite);
 | 
						|
	igt_assert(ret == 0 || errno == EINVAL);
 | 
						|
	gem_close(fd, handle);
 | 
						|
	free(ptr);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static unsigned char counter;
 | 
						|
 | 
						|
static void (* volatile orig_sigbus)(int sig, siginfo_t *info, void *param);
 | 
						|
static volatile unsigned long sigbus_start;
 | 
						|
static volatile long sigbus_cnt = -1;
 | 
						|
 | 
						|
static void *umap(int fd, uint32_t handle)
 | 
						|
{
 | 
						|
	void *ptr;
 | 
						|
 | 
						|
	if (gem_has_llc(fd)) {
 | 
						|
		ptr = gem_mmap__gtt(fd, handle, sizeof(linear),
 | 
						|
				    PROT_READ | PROT_WRITE);
 | 
						|
	} else {
 | 
						|
		uint32_t tmp = gem_create(fd, sizeof(linear));
 | 
						|
		copy(fd, tmp, handle, 0);
 | 
						|
		ptr = gem_mmap__cpu(fd, tmp, 0, sizeof(linear), PROT_READ);
 | 
						|
		gem_close(fd, tmp);
 | 
						|
	}
 | 
						|
 | 
						|
	return ptr;
 | 
						|
}
 | 
						|
 | 
						|
static void
 | 
						|
check_bo(int fd1, uint32_t handle1, int is_userptr, int fd2, uint32_t handle2)
 | 
						|
{
 | 
						|
	unsigned char *ptr1, *ptr2;
 | 
						|
	unsigned long size = sizeof(linear);
 | 
						|
 | 
						|
	ptr2 = umap(fd2, handle2);
 | 
						|
	if (is_userptr)
 | 
						|
		ptr1 = is_userptr > 0 ? get_handle_ptr(handle1) : ptr2;
 | 
						|
	else
 | 
						|
		ptr1 = umap(fd1, handle1);
 | 
						|
 | 
						|
	igt_assert(ptr1);
 | 
						|
	igt_assert(ptr2);
 | 
						|
 | 
						|
	sigbus_start = (unsigned long)ptr2;
 | 
						|
	igt_assert(memcmp(ptr1, ptr2, sizeof(linear)) == 0);
 | 
						|
 | 
						|
	if (gem_has_llc(fd1)) {
 | 
						|
		counter++;
 | 
						|
		memset(ptr1, counter, size);
 | 
						|
		memset(ptr2, counter, size);
 | 
						|
	}
 | 
						|
 | 
						|
	if (!is_userptr)
 | 
						|
		munmap(ptr1, sizeof(linear));
 | 
						|
	munmap(ptr2, sizeof(linear));
 | 
						|
}
 | 
						|
 | 
						|
static int export_handle(int fd, uint32_t handle, int *outfd)
 | 
						|
{
 | 
						|
	struct drm_prime_handle args;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	args.handle = handle;
 | 
						|
	args.flags = DRM_CLOEXEC;
 | 
						|
	args.fd = -1;
 | 
						|
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args);
 | 
						|
	if (ret)
 | 
						|
		ret = errno;
 | 
						|
	*outfd = args.fd;
 | 
						|
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static void sigbus(int sig, siginfo_t *info, void *param)
 | 
						|
{
 | 
						|
	unsigned long ptr = (unsigned long)info->si_addr;
 | 
						|
	void *addr;
 | 
						|
 | 
						|
	if (ptr >= sigbus_start &&
 | 
						|
	    ptr < sigbus_start + sizeof(linear)) {
 | 
						|
		/* replace mapping to allow progress */
 | 
						|
		munmap((void *)sigbus_start, sizeof(linear));
 | 
						|
		addr = mmap((void *)sigbus_start, sizeof(linear),
 | 
						|
			    PROT_READ | PROT_WRITE,
 | 
						|
			    MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0);
 | 
						|
		igt_assert((unsigned long)addr == sigbus_start);
 | 
						|
		memset(addr, counter, sizeof(linear));
 | 
						|
 | 
						|
		sigbus_cnt++;
 | 
						|
		return;
 | 
						|
	}
 | 
						|
 | 
						|
	if (orig_sigbus)
 | 
						|
		orig_sigbus(sig, info, param);
 | 
						|
	igt_assert(0);
 | 
						|
}
 | 
						|
 | 
						|
static int test_dmabuf(void)
 | 
						|
{
 | 
						|
	int fd1, fd2;
 | 
						|
	uint32_t handle, handle_import;
 | 
						|
	int dma_buf_fd = -1;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	fd1 = drm_open_driver(DRIVER_INTEL);
 | 
						|
 | 
						|
	handle = create_userptr_bo(fd1, sizeof(linear));
 | 
						|
	memset(get_handle_ptr(handle), counter, sizeof(linear));
 | 
						|
 | 
						|
	ret = export_handle(fd1, handle, &dma_buf_fd);
 | 
						|
	if (userptr_flags & LOCAL_I915_USERPTR_UNSYNCHRONIZED && ret) {
 | 
						|
		igt_assert(ret == EINVAL || ret == ENODEV);
 | 
						|
		free_userptr_bo(fd1, handle);
 | 
						|
		close(fd1);
 | 
						|
		return 0;
 | 
						|
	} else {
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
		igt_assert_lte(0, dma_buf_fd);
 | 
						|
	}
 | 
						|
 | 
						|
	fd2 = drm_open_driver(DRIVER_INTEL);
 | 
						|
	handle_import = prime_fd_to_handle(fd2, dma_buf_fd);
 | 
						|
	check_bo(fd1, handle, 1, fd2, handle_import);
 | 
						|
 | 
						|
	/* close dma_buf, check whether nothing disappears. */
 | 
						|
	close(dma_buf_fd);
 | 
						|
	check_bo(fd1, handle, 1, fd2, handle_import);
 | 
						|
 | 
						|
	/* destroy userptr object and expect SIGBUS */
 | 
						|
	free_userptr_bo(fd1, handle);
 | 
						|
	close(fd1);
 | 
						|
 | 
						|
	if (gem_has_llc(fd2)) {
 | 
						|
		struct sigaction sigact, orig_sigact;
 | 
						|
 | 
						|
		memset(&sigact, 0, sizeof(sigact));
 | 
						|
		sigact.sa_sigaction = sigbus;
 | 
						|
		sigact.sa_flags = SA_SIGINFO;
 | 
						|
		ret = sigaction(SIGBUS, &sigact, &orig_sigact);
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
		orig_sigbus = orig_sigact.sa_sigaction;
 | 
						|
 | 
						|
		sigbus_cnt = 0;
 | 
						|
		check_bo(fd2, handle_import, -1, fd2, handle_import);
 | 
						|
		igt_assert(sigbus_cnt > 0);
 | 
						|
 | 
						|
		ret = sigaction(SIGBUS, &orig_sigact, NULL);
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
	}
 | 
						|
 | 
						|
	close(fd2);
 | 
						|
	reset_handle_ptr();
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_usage_restrictions(int fd)
 | 
						|
{
 | 
						|
	void *ptr;
 | 
						|
	int ret;
 | 
						|
	uint32_t handle;
 | 
						|
 | 
						|
	igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE * 2) == 0);
 | 
						|
 | 
						|
	/* Address not aligned. */
 | 
						|
	ret = gem_userptr(fd, (char *)ptr + 1, PAGE_SIZE, 0, &handle);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	/* Size not rounded to page size. */
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE - 1, 0, &handle);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	/* Both wrong. */
 | 
						|
	ret = gem_userptr(fd, (char *)ptr + 1, PAGE_SIZE - 1, 0, &handle);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	/* Read-only not supported. */
 | 
						|
	ret = gem_userptr(fd, (char *)ptr, PAGE_SIZE, 1, &handle);
 | 
						|
	igt_assert_neq(ret, 0);
 | 
						|
 | 
						|
	free(ptr);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_create_destroy(int fd, int time)
 | 
						|
{
 | 
						|
	struct timespec start, now;
 | 
						|
	uint32_t handle;
 | 
						|
	void *ptr;
 | 
						|
	int n;
 | 
						|
 | 
						|
	igt_fork_signal_helper();
 | 
						|
 | 
						|
	clock_gettime(CLOCK_MONOTONIC, &start);
 | 
						|
	do {
 | 
						|
		for (n = 0; n < 1000; n++) {
 | 
						|
			igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
 | 
						|
			do_or_die(gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle));
 | 
						|
 | 
						|
			gem_close(fd, handle);
 | 
						|
			free(ptr);
 | 
						|
		}
 | 
						|
 | 
						|
		clock_gettime(CLOCK_MONOTONIC, &now);
 | 
						|
		now.tv_sec -= time;
 | 
						|
	} while (now.tv_sec < start.tv_sec ||
 | 
						|
		 (now.tv_sec == start.tv_sec && now.tv_nsec < start.tv_nsec));
 | 
						|
 | 
						|
	igt_stop_signal_helper();
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int test_coherency(int fd, int count)
 | 
						|
{
 | 
						|
	uint32_t *memory;
 | 
						|
	uint32_t *cpu, *cpu_val;
 | 
						|
	uint32_t *gpu, *gpu_val;
 | 
						|
	uint32_t start = 0;
 | 
						|
	int i, ret;
 | 
						|
 | 
						|
	igt_info("Using 2x%d 1MiB buffers\n", count);
 | 
						|
	intel_require_memory(2*count, sizeof(linear), CHECK_RAM);
 | 
						|
 | 
						|
	ret = posix_memalign((void **)&memory, PAGE_SIZE, count*sizeof(linear));
 | 
						|
	igt_assert(ret == 0 && memory);
 | 
						|
 | 
						|
	gpu = malloc(sizeof(uint32_t)*count*4);
 | 
						|
	gpu_val = gpu + count;
 | 
						|
	cpu = gpu_val + count;
 | 
						|
	cpu_val = cpu + count;
 | 
						|
 | 
						|
	for (i = 0; i < count; i++) {
 | 
						|
		gpu[i] = create_bo(fd, start);
 | 
						|
		gpu_val[i] = start;
 | 
						|
		start += WIDTH*HEIGHT;
 | 
						|
	}
 | 
						|
 | 
						|
	for (i = 0; i < count; i++) {
 | 
						|
		cpu[i] = create_userptr(fd, start, memory+i*WIDTH*HEIGHT);
 | 
						|
		cpu_val[i] = start;
 | 
						|
		start += WIDTH*HEIGHT;
 | 
						|
	}
 | 
						|
 | 
						|
	igt_info("Verifying initialisation...\n");
 | 
						|
	for (i = 0; i < count; i++) {
 | 
						|
		check_gpu(fd, gpu[i], gpu_val[i]);
 | 
						|
		check_cpu(memory+i*WIDTH*HEIGHT, cpu_val[i]);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_info("Cyclic blits cpu->gpu, forward...\n");
 | 
						|
	for (i = 0; i < count * 4; i++) {
 | 
						|
		int src = i % count;
 | 
						|
		int dst = (i + 1) % count;
 | 
						|
 | 
						|
		copy(fd, gpu[dst], cpu[src], 0);
 | 
						|
		gpu_val[dst] = cpu_val[src];
 | 
						|
	}
 | 
						|
	for (i = 0; i < count; i++)
 | 
						|
		check_gpu(fd, gpu[i], gpu_val[i]);
 | 
						|
 | 
						|
	igt_info("Cyclic blits gpu->cpu, backward...\n");
 | 
						|
	for (i = 0; i < count * 4; i++) {
 | 
						|
		int src = (i + 1) % count;
 | 
						|
		int dst = i % count;
 | 
						|
 | 
						|
		copy(fd, cpu[dst], gpu[src], 0);
 | 
						|
		cpu_val[dst] = gpu_val[src];
 | 
						|
	}
 | 
						|
	for (i = 0; i < count; i++) {
 | 
						|
		gem_userptr_sync(fd, cpu[i]);
 | 
						|
		check_cpu(memory+i*WIDTH*HEIGHT, cpu_val[i]);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_info("Random blits...\n");
 | 
						|
	for (i = 0; i < count * 4; i++) {
 | 
						|
		int src = random() % count;
 | 
						|
		int dst = random() % count;
 | 
						|
 | 
						|
		if (random() & 1) {
 | 
						|
			copy(fd, gpu[dst], cpu[src], 0);
 | 
						|
			gpu_val[dst] = cpu_val[src];
 | 
						|
		} else {
 | 
						|
			copy(fd, cpu[dst], gpu[src], 0);
 | 
						|
			cpu_val[dst] = gpu_val[src];
 | 
						|
		}
 | 
						|
	}
 | 
						|
	for (i = 0; i < count; i++) {
 | 
						|
		check_gpu(fd, gpu[i], gpu_val[i]);
 | 
						|
		gem_close(fd, gpu[i]);
 | 
						|
 | 
						|
		gem_userptr_sync(fd, cpu[i]);
 | 
						|
		check_cpu(memory+i*WIDTH*HEIGHT, cpu_val[i]);
 | 
						|
		gem_close(fd, cpu[i]);
 | 
						|
	}
 | 
						|
 | 
						|
	free(gpu);
 | 
						|
	free(memory);
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct igt_eviction_test_ops fault_ops = {
 | 
						|
	.create = create_userptr_bo,
 | 
						|
	.flink = flink_userptr_bo,
 | 
						|
	.close = free_userptr_bo,
 | 
						|
	.copy = blit,
 | 
						|
	.clear = clear,
 | 
						|
};
 | 
						|
 | 
						|
static int can_swap(void)
 | 
						|
{
 | 
						|
	unsigned long as, ram;
 | 
						|
 | 
						|
	/* Cannot swap if not enough address space */
 | 
						|
 | 
						|
	/* FIXME: Improve check criteria. */
 | 
						|
	if (sizeof(void*) < 8)
 | 
						|
		as = 3 * 1024;
 | 
						|
	else
 | 
						|
		as = 256 * 1024; /* Just a big number */
 | 
						|
 | 
						|
	ram = intel_get_total_ram_mb();
 | 
						|
 | 
						|
	if ((as - 128) < (ram - 256))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	return 1;
 | 
						|
}
 | 
						|
 | 
						|
static void test_forking_evictions(int fd, int size, int count,
 | 
						|
			     unsigned flags)
 | 
						|
{
 | 
						|
	int trash_count;
 | 
						|
	int num_threads;
 | 
						|
 | 
						|
	trash_count = intel_get_total_ram_mb() * 11 / 10;
 | 
						|
	/* Use the fact test will spawn a number of child
 | 
						|
	 * processes meaning swapping will be triggered system
 | 
						|
	 * wide even if one process on it's own can't do it.
 | 
						|
	 */
 | 
						|
	num_threads = min(sysconf(_SC_NPROCESSORS_ONLN) * 4, 12);
 | 
						|
	trash_count /= num_threads;
 | 
						|
	if (count > trash_count)
 | 
						|
		count = trash_count;
 | 
						|
 | 
						|
	forking_evictions(fd, &fault_ops, size, count, trash_count, flags);
 | 
						|
	reset_handle_ptr();
 | 
						|
}
 | 
						|
 | 
						|
static void test_mlocked_evictions(int fd, int size, int count)
 | 
						|
{
 | 
						|
	mlocked_evictions(fd, &fault_ops, size, count);
 | 
						|
	reset_handle_ptr();
 | 
						|
}
 | 
						|
 | 
						|
static void test_swapping_evictions(int fd, int size, int count)
 | 
						|
{
 | 
						|
	int trash_count;
 | 
						|
 | 
						|
	igt_skip_on_f(!can_swap(),
 | 
						|
		"Not enough process address space for swapping tests.\n");
 | 
						|
 | 
						|
	trash_count = intel_get_total_ram_mb() * 11 / 10;
 | 
						|
 | 
						|
	swapping_evictions(fd, &fault_ops, size, count, trash_count);
 | 
						|
	reset_handle_ptr();
 | 
						|
}
 | 
						|
 | 
						|
static void test_minor_evictions(int fd, int size, int count)
 | 
						|
{
 | 
						|
	minor_evictions(fd, &fault_ops, size, count);
 | 
						|
	reset_handle_ptr();
 | 
						|
}
 | 
						|
 | 
						|
static void test_major_evictions(int fd, int size, int count)
 | 
						|
{
 | 
						|
	major_evictions(fd, &fault_ops, size, count);
 | 
						|
	reset_handle_ptr();
 | 
						|
}
 | 
						|
 | 
						|
static void test_overlap(int fd, int expected)
 | 
						|
{
 | 
						|
	char *ptr;
 | 
						|
	int ret;
 | 
						|
	uint32_t handle, handle2;
 | 
						|
 | 
						|
	igt_assert(posix_memalign((void *)&ptr, PAGE_SIZE, PAGE_SIZE * 3) == 0);
 | 
						|
 | 
						|
	ret = gem_userptr(fd, ptr + PAGE_SIZE, PAGE_SIZE, 0, &handle);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	/* before, no overlap */
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	/* after, no overlap */
 | 
						|
	ret = gem_userptr(fd, ptr + PAGE_SIZE * 2, PAGE_SIZE, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	/* exactly overlapping */
 | 
						|
	ret = gem_userptr(fd, ptr + PAGE_SIZE, PAGE_SIZE, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert(ret == 0 || ret == expected);
 | 
						|
 | 
						|
	/* start overlaps */
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE * 2, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert(ret == 0 || ret == expected);
 | 
						|
 | 
						|
	/* end overlaps */
 | 
						|
	ret = gem_userptr(fd, ptr + PAGE_SIZE, PAGE_SIZE * 2, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert(ret == 0 || ret == expected);
 | 
						|
 | 
						|
	/* subsumes */
 | 
						|
	ret = gem_userptr(fd, ptr, PAGE_SIZE * 3, 0, &handle2);
 | 
						|
	if (ret == 0)
 | 
						|
		gem_close(fd, handle2);
 | 
						|
	igt_assert(ret == 0 || ret == expected);
 | 
						|
 | 
						|
	gem_close(fd, handle);
 | 
						|
	free(ptr);
 | 
						|
}
 | 
						|
 | 
						|
static void test_unmap(int fd, int expected)
 | 
						|
{
 | 
						|
	char *ptr, *bo_ptr;
 | 
						|
	const unsigned int num_obj = 3;
 | 
						|
	unsigned int i;
 | 
						|
	uint32_t bo[num_obj + 1];
 | 
						|
	size_t map_size = sizeof(linear) * num_obj + (PAGE_SIZE - 1);
 | 
						|
	int ret;
 | 
						|
 | 
						|
	ptr = mmap(NULL, map_size, PROT_READ | PROT_WRITE,
 | 
						|
				MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
 | 
						|
	igt_assert(ptr != MAP_FAILED);
 | 
						|
 | 
						|
	bo_ptr = (char *)ALIGN((unsigned long)ptr, PAGE_SIZE);
 | 
						|
 | 
						|
	for (i = 0; i < num_obj; i++, bo_ptr += sizeof(linear)) {
 | 
						|
		ret = gem_userptr(fd, bo_ptr, sizeof(linear), 0, &bo[i]);
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
	}
 | 
						|
 | 
						|
	bo[num_obj] = create_bo(fd, 0);
 | 
						|
 | 
						|
	for (i = 0; i < num_obj; i++)
 | 
						|
		copy(fd, bo[num_obj], bo[i], 0);
 | 
						|
 | 
						|
	ret = munmap(ptr, map_size);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	for (i = 0; i < num_obj; i++)
 | 
						|
		copy(fd, bo[num_obj], bo[i], expected);
 | 
						|
 | 
						|
	for (i = 0; i < (num_obj + 1); i++)
 | 
						|
		gem_close(fd, bo[i]);
 | 
						|
}
 | 
						|
 | 
						|
static void test_unmap_after_close(int fd)
 | 
						|
{
 | 
						|
	char *ptr, *bo_ptr;
 | 
						|
	const unsigned int num_obj = 3;
 | 
						|
	unsigned int i;
 | 
						|
	uint32_t bo[num_obj + 1];
 | 
						|
	size_t map_size = sizeof(linear) * num_obj + (PAGE_SIZE - 1);
 | 
						|
	int ret;
 | 
						|
 | 
						|
	ptr = mmap(NULL, map_size, PROT_READ | PROT_WRITE,
 | 
						|
				MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
 | 
						|
	igt_assert(ptr != MAP_FAILED);
 | 
						|
 | 
						|
	bo_ptr = (char *)ALIGN((unsigned long)ptr, PAGE_SIZE);
 | 
						|
 | 
						|
	for (i = 0; i < num_obj; i++, bo_ptr += sizeof(linear)) {
 | 
						|
		ret = gem_userptr(fd, bo_ptr, sizeof(linear), 0, &bo[i]);
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
	}
 | 
						|
 | 
						|
	bo[num_obj] = create_bo(fd, 0);
 | 
						|
 | 
						|
	for (i = 0; i < num_obj; i++)
 | 
						|
		copy(fd, bo[num_obj], bo[i], 0);
 | 
						|
 | 
						|
	for (i = 0; i < (num_obj + 1); i++)
 | 
						|
		gem_close(fd, bo[i]);
 | 
						|
 | 
						|
	ret = munmap(ptr, map_size);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
}
 | 
						|
 | 
						|
static void test_unmap_cycles(int fd, int expected)
 | 
						|
{
 | 
						|
	int i;
 | 
						|
 | 
						|
	for (i = 0; i < 1000; i++)
 | 
						|
		test_unmap(fd, expected);
 | 
						|
}
 | 
						|
 | 
						|
#define MM_STRESS_LOOPS 100000
 | 
						|
 | 
						|
struct stress_thread_data {
 | 
						|
	unsigned int stop;
 | 
						|
	int exit_code;
 | 
						|
};
 | 
						|
 | 
						|
static void *mm_stress_thread(void *data)
 | 
						|
{
 | 
						|
	struct stress_thread_data *stdata = (struct stress_thread_data *)data;
 | 
						|
	void *ptr;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	while (!stdata->stop) {
 | 
						|
		ptr = mmap(NULL, PAGE_SIZE, PROT_READ | PROT_WRITE,
 | 
						|
				MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
 | 
						|
		if (ptr == MAP_FAILED) {
 | 
						|
			stdata->exit_code = -EFAULT;
 | 
						|
			break;
 | 
						|
		}
 | 
						|
		ret = munmap(ptr, PAGE_SIZE);
 | 
						|
		if (ret) {
 | 
						|
		        stdata->exit_code = errno;
 | 
						|
		        break;
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
 | 
						|
static void test_stress_mm(int fd)
 | 
						|
{
 | 
						|
	int ret;
 | 
						|
	pthread_t t;
 | 
						|
	unsigned int loops = MM_STRESS_LOOPS;
 | 
						|
	uint32_t handle;
 | 
						|
	void *ptr;
 | 
						|
	struct stress_thread_data stdata;
 | 
						|
 | 
						|
	memset(&stdata, 0, sizeof(stdata));
 | 
						|
 | 
						|
	igt_assert(posix_memalign(&ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
 | 
						|
	ret = pthread_create(&t, NULL, mm_stress_thread, &stdata);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	while (loops--) {
 | 
						|
		ret = gem_userptr(fd, ptr, PAGE_SIZE, 0, &handle);
 | 
						|
		igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
		gem_close(fd, handle);
 | 
						|
	}
 | 
						|
 | 
						|
	free(ptr);
 | 
						|
 | 
						|
	stdata.stop = 1;
 | 
						|
	ret = pthread_join(t, NULL);
 | 
						|
	igt_assert_eq(ret, 0);
 | 
						|
 | 
						|
	igt_assert_eq(stdata.exit_code, 0);
 | 
						|
}
 | 
						|
 | 
						|
struct userptr_close_thread_data {
 | 
						|
	int fd;
 | 
						|
	void *ptr;
 | 
						|
	bool overlap;
 | 
						|
	bool stop;
 | 
						|
	pthread_mutex_t mutex;
 | 
						|
};
 | 
						|
 | 
						|
static void *mm_userptr_close_thread(void *data)
 | 
						|
{
 | 
						|
	struct userptr_close_thread_data *t = (struct userptr_close_thread_data *)data;
 | 
						|
	int num_handles = t->overlap ? 2 : 1;
 | 
						|
 | 
						|
	uint32_t handle[num_handles];
 | 
						|
 | 
						|
	/* Be pedantic and enforce the required memory barriers */
 | 
						|
	pthread_mutex_lock(&t->mutex);
 | 
						|
	while (!t->stop) {
 | 
						|
		pthread_mutex_unlock(&t->mutex);
 | 
						|
		for (int i = 0; i < num_handles; i++)
 | 
						|
			igt_assert_eq(gem_userptr(t->fd, t->ptr, PAGE_SIZE, 0, &handle[i]),
 | 
						|
				      0);
 | 
						|
		for (int i = 0; i < num_handles; i++)
 | 
						|
			gem_close(t->fd, handle[i]);
 | 
						|
		pthread_mutex_lock(&t->mutex);
 | 
						|
	}
 | 
						|
	pthread_mutex_unlock(&t->mutex);
 | 
						|
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
 | 
						|
static void test_invalidate_close_race(int fd, bool overlap)
 | 
						|
{
 | 
						|
	pthread_t t;
 | 
						|
	unsigned int loops = MM_STRESS_LOOPS;
 | 
						|
	struct userptr_close_thread_data t_data;
 | 
						|
 | 
						|
	memset(&t_data, 0, sizeof(t_data));
 | 
						|
	t_data.fd = fd;
 | 
						|
	t_data.overlap = overlap;
 | 
						|
	igt_assert(posix_memalign(&t_data.ptr, PAGE_SIZE, PAGE_SIZE) == 0);
 | 
						|
	pthread_mutex_init(&t_data.mutex, NULL);
 | 
						|
 | 
						|
	igt_assert(pthread_create(&t, NULL, mm_userptr_close_thread, &t_data) == 0);
 | 
						|
 | 
						|
	while (loops--) {
 | 
						|
		mprotect(t_data.ptr, PAGE_SIZE, PROT_READ | PROT_WRITE | PROT_EXEC);
 | 
						|
		mprotect(t_data.ptr, PAGE_SIZE, PROT_READ | PROT_WRITE);
 | 
						|
	}
 | 
						|
 | 
						|
	pthread_mutex_lock(&t_data.mutex);
 | 
						|
	t_data.stop = 1;
 | 
						|
	pthread_mutex_unlock(&t_data.mutex);
 | 
						|
 | 
						|
	pthread_join(t, NULL);
 | 
						|
 | 
						|
	pthread_mutex_destroy(&t_data.mutex);
 | 
						|
	free(t_data.ptr);
 | 
						|
}
 | 
						|
 | 
						|
unsigned int total_ram;
 | 
						|
uint64_t aperture_size;
 | 
						|
int fd, count;
 | 
						|
 | 
						|
 | 
						|
int main(int argc, char **argv)
 | 
						|
{
 | 
						|
	int size = sizeof(linear);
 | 
						|
 | 
						|
	igt_subtest_init(argc, argv);
 | 
						|
 | 
						|
	igt_fixture {
 | 
						|
		int ret;
 | 
						|
 | 
						|
		fd = drm_open_driver(DRIVER_INTEL);
 | 
						|
		igt_assert(fd >= 0);
 | 
						|
 | 
						|
		ret = has_userptr(fd);
 | 
						|
		igt_skip_on_f(ret == 0, "No userptr support - %s (%d)\n",
 | 
						|
			      strerror(errno), ret);
 | 
						|
 | 
						|
		size = sizeof(linear);
 | 
						|
 | 
						|
		aperture_size = gem_aperture_size(fd);
 | 
						|
		igt_info("Aperture size is %lu MiB\n", (long)(aperture_size / (1024*1024)));
 | 
						|
 | 
						|
		if (argc > 1)
 | 
						|
			count = atoi(argv[1]);
 | 
						|
		if (count == 0)
 | 
						|
			count = 2 * aperture_size / (1024*1024) / 3;
 | 
						|
 | 
						|
		total_ram = intel_get_total_ram_mb();
 | 
						|
		igt_info("Total RAM is %u MiB\n", total_ram);
 | 
						|
 | 
						|
		if (count > total_ram * 3 / 4) {
 | 
						|
			count = intel_get_total_ram_mb() * 3 / 4;
 | 
						|
			igt_info("Not enough RAM to run test, reducing buffer count.\n");
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	igt_subtest("input-checking")
 | 
						|
		test_input_checking(fd);
 | 
						|
 | 
						|
	igt_subtest("usage-restrictions")
 | 
						|
		test_usage_restrictions(fd);
 | 
						|
 | 
						|
	igt_subtest("invalid-null-pointer")
 | 
						|
		test_invalid_null_pointer(fd);
 | 
						|
 | 
						|
	igt_subtest("invalid-gtt-mapping")
 | 
						|
		test_invalid_gtt_mapping(fd);
 | 
						|
 | 
						|
	igt_subtest("forked-access")
 | 
						|
		test_forked_access(fd);
 | 
						|
 | 
						|
	igt_subtest("forbidden-operations")
 | 
						|
		test_forbidden_ops(fd);
 | 
						|
 | 
						|
	igt_info("Testing unsynchronized mappings...\n");
 | 
						|
	gem_userptr_test_unsynchronized();
 | 
						|
 | 
						|
	igt_subtest("create-destroy-unsync")
 | 
						|
		test_create_destroy(fd, 5);
 | 
						|
 | 
						|
	igt_subtest("unsync-overlap")
 | 
						|
		test_overlap(fd, 0);
 | 
						|
 | 
						|
	igt_subtest("unsync-unmap")
 | 
						|
		test_unmap(fd, 0);
 | 
						|
 | 
						|
	igt_subtest("unsync-unmap-cycles")
 | 
						|
		test_unmap_cycles(fd, 0);
 | 
						|
 | 
						|
	igt_subtest("unsync-unmap-after-close")
 | 
						|
		test_unmap_after_close(fd);
 | 
						|
 | 
						|
	igt_subtest("coherency-unsync")
 | 
						|
		test_coherency(fd, count);
 | 
						|
 | 
						|
	igt_subtest("dmabuf-unsync")
 | 
						|
		test_dmabuf();
 | 
						|
 | 
						|
	for (unsigned flags = 0; flags < ALL_FORKING_EVICTIONS + 1; flags++) {
 | 
						|
		igt_subtest_f("forked-unsync%s%s%s-%s",
 | 
						|
		    flags & FORKING_EVICTIONS_SWAPPING ? "-swapping" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_DUP_DRMFD ? "-multifd" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_MEMORY_PRESSURE ?
 | 
						|
				"-mempressure" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_INTERRUPTIBLE ?
 | 
						|
				"interruptible" : "normal") {
 | 
						|
			test_forking_evictions(fd, size, count, flags);
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	igt_subtest("mlocked-unsync-normal")
 | 
						|
		test_mlocked_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("swapping-unsync-normal")
 | 
						|
		test_swapping_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("minor-unsync-normal")
 | 
						|
		test_minor_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("major-unsync-normal") {
 | 
						|
		size = 200 * 1024 * 1024;
 | 
						|
		count = (gem_aperture_size(fd) / size) + 2;
 | 
						|
		test_major_evictions(fd, size, count);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_fixture {
 | 
						|
		size = sizeof(linear);
 | 
						|
		count = 2 * gem_aperture_size(fd) / (1024*1024) / 3;
 | 
						|
		if (count > total_ram * 3 / 4)
 | 
						|
			count = intel_get_total_ram_mb() * 3 / 4;
 | 
						|
	}
 | 
						|
 | 
						|
	igt_fork_signal_helper();
 | 
						|
 | 
						|
	igt_subtest("mlocked-unsync-interruptible")
 | 
						|
		test_mlocked_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("swapping-unsync-interruptible")
 | 
						|
		test_swapping_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("minor-unsync-interruptible")
 | 
						|
		test_minor_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("major-unsync-interruptible") {
 | 
						|
		size = 200 * 1024 * 1024;
 | 
						|
		count = (gem_aperture_size(fd) / size) + 2;
 | 
						|
		test_major_evictions(fd, size, count);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_stop_signal_helper();
 | 
						|
 | 
						|
	igt_info("Testing synchronized mappings...\n");
 | 
						|
 | 
						|
	igt_fixture {
 | 
						|
		size = sizeof(linear);
 | 
						|
		count = 2 * gem_aperture_size(fd) / (1024*1024) / 3;
 | 
						|
		if (count > total_ram * 3 / 4)
 | 
						|
			count = intel_get_total_ram_mb() * 3 / 4;
 | 
						|
	}
 | 
						|
 | 
						|
	gem_userptr_test_synchronized();
 | 
						|
 | 
						|
	igt_subtest("process-exit")
 | 
						|
		test_process_exit(fd, 0);
 | 
						|
 | 
						|
	igt_subtest("process-exit-gtt")
 | 
						|
		test_process_exit(fd, PE_GTT_MAP);
 | 
						|
 | 
						|
	igt_subtest("process-exit-busy")
 | 
						|
		test_process_exit(fd, PE_BUSY);
 | 
						|
 | 
						|
	igt_subtest("process-exit-gtt-busy")
 | 
						|
		test_process_exit(fd, PE_GTT_MAP | PE_BUSY);
 | 
						|
 | 
						|
	igt_subtest("create-destroy-sync")
 | 
						|
		test_create_destroy(fd, 5);
 | 
						|
 | 
						|
	igt_subtest("sync-overlap")
 | 
						|
		test_overlap(fd, EINVAL);
 | 
						|
 | 
						|
	igt_subtest("sync-unmap")
 | 
						|
		test_unmap(fd, EFAULT);
 | 
						|
 | 
						|
	igt_subtest("sync-unmap-cycles")
 | 
						|
		test_unmap_cycles(fd, EFAULT);
 | 
						|
 | 
						|
	igt_subtest("sync-unmap-after-close")
 | 
						|
		test_unmap_after_close(fd);
 | 
						|
 | 
						|
	igt_subtest("stress-mm")
 | 
						|
		test_stress_mm(fd);
 | 
						|
 | 
						|
	igt_subtest("stress-mm-invalidate-close")
 | 
						|
		test_invalidate_close_race(fd, false);
 | 
						|
 | 
						|
	igt_subtest("stress-mm-invalidate-close-overlap")
 | 
						|
		test_invalidate_close_race(fd, true);
 | 
						|
 | 
						|
	igt_subtest("coherency-sync")
 | 
						|
		test_coherency(fd, count);
 | 
						|
 | 
						|
	igt_subtest("dmabuf-sync")
 | 
						|
		test_dmabuf();
 | 
						|
 | 
						|
	for (unsigned flags = 0; flags < ALL_FORKING_EVICTIONS + 1; flags++) {
 | 
						|
		igt_subtest_f("forked-sync%s%s%s-%s",
 | 
						|
		    flags & FORKING_EVICTIONS_SWAPPING ? "-swapping" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_DUP_DRMFD ? "-multifd" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_MEMORY_PRESSURE ?
 | 
						|
				"-mempressure" : "",
 | 
						|
		    flags & FORKING_EVICTIONS_INTERRUPTIBLE ?
 | 
						|
				"interruptible" : "normal") {
 | 
						|
			test_forking_evictions(fd, size, count, flags);
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	igt_subtest("mlocked-normal-sync")
 | 
						|
		test_mlocked_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("swapping-normal-sync")
 | 
						|
		test_swapping_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("minor-normal-sync")
 | 
						|
		test_minor_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("major-normal-sync") {
 | 
						|
		size = 200 * 1024 * 1024;
 | 
						|
		count = (gem_aperture_size(fd) / size) + 2;
 | 
						|
		test_major_evictions(fd, size, count);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_fixture {
 | 
						|
		size = 1024 * 1024;
 | 
						|
		count = 2 * gem_aperture_size(fd) / (1024*1024) / 3;
 | 
						|
		if (count > total_ram * 3 / 4)
 | 
						|
			count = intel_get_total_ram_mb() * 3 / 4;
 | 
						|
	}
 | 
						|
 | 
						|
	igt_fork_signal_helper();
 | 
						|
 | 
						|
	igt_subtest("mlocked-sync-interruptible")
 | 
						|
		test_mlocked_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("swapping-sync-interruptible")
 | 
						|
		test_swapping_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("minor-sync-interruptible")
 | 
						|
		test_minor_evictions(fd, size, count);
 | 
						|
 | 
						|
	igt_subtest("major-sync-interruptible") {
 | 
						|
		size = 200 * 1024 * 1024;
 | 
						|
		count = (gem_aperture_size(fd) / size) + 2;
 | 
						|
		test_major_evictions(fd, size, count);
 | 
						|
	}
 | 
						|
 | 
						|
	igt_stop_signal_helper();
 | 
						|
 | 
						|
	igt_subtest("access-control")
 | 
						|
		test_access_control(fd);
 | 
						|
 | 
						|
	igt_exit();
 | 
						|
}
 |