mirror of
				https://github.com/tiagovignatti/intel-gpu-tools.git
				synced 2025-11-04 12:07:12 +00:00 
			
		
		
		
	Apply the new API to all call sites within the test suite using the following semantic patch: // Semantic patch for replacing drm_open_any* with arch-specific drm_open_driver* calls @@ identifier i =~ "\bdrm_open_any\b"; @@ - i() + drm_open_driver(DRIVER_INTEL) @@ identifier i =~ "\bdrm_open_any_master\b"; @@ - i() + drm_open_driver_master(DRIVER_INTEL) @@ identifier i =~ "\bdrm_open_any_render\b"; @@ - i() + drm_open_driver_render(DRIVER_INTEL) @@ identifier i =~ "\b__drm_open_any\b"; @@ - i() + __drm_open_driver(DRIVER_INTEL) Signed-off-by: Micah Fedke <micah.fedke@collabora.co.uk> Signed-off-by: Thomas Wood <thomas.wood@intel.com>
		
			
				
	
	
		
			277 lines
		
	
	
		
			7.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			277 lines
		
	
	
		
			7.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright © 2012 Intel Corporation
 | 
						|
 *
 | 
						|
 * Permission is hereby granted, free of charge, to any person obtaining a
 | 
						|
 * copy of this software and associated documentation files (the "Software"),
 | 
						|
 * to deal in the Software without restriction, including without limitation
 | 
						|
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 | 
						|
 * and/or sell copies of the Software, and to permit persons to whom the
 | 
						|
 * Software is furnished to do so, subject to the following conditions:
 | 
						|
 *
 | 
						|
 * The above copyright notice and this permission notice (including the next
 | 
						|
 * paragraph) shall be included in all copies or substantial portions of the
 | 
						|
 * Software.
 | 
						|
 *
 | 
						|
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | 
						|
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | 
						|
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 | 
						|
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | 
						|
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 | 
						|
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 | 
						|
 * IN THE SOFTWARE.
 | 
						|
 *
 | 
						|
 * Authors:
 | 
						|
 *    Ben Widawsky <ben@bwidawsk.net>
 | 
						|
 *
 | 
						|
 */
 | 
						|
 | 
						|
#include "igt.h"
 | 
						|
#include <stdio.h>
 | 
						|
#include <time.h>
 | 
						|
#include <stdlib.h>
 | 
						|
#include <sys/ioctl.h>
 | 
						|
#include <stdio.h>
 | 
						|
#include <string.h>
 | 
						|
#include <fcntl.h>
 | 
						|
#include <inttypes.h>
 | 
						|
#include <errno.h>
 | 
						|
#include <sys/stat.h>
 | 
						|
#include <sys/time.h>
 | 
						|
 | 
						|
#include <drm.h>
 | 
						|
 | 
						|
#include "intel_bufmgr.h"
 | 
						|
 | 
						|
#define MSEC_PER_SEC	1000L
 | 
						|
#define USEC_PER_MSEC	1000L
 | 
						|
#define NSEC_PER_USEC	1000L
 | 
						|
#define NSEC_PER_MSEC	1000000L
 | 
						|
#define USEC_PER_SEC	1000000L
 | 
						|
#define NSEC_PER_SEC	1000000000L
 | 
						|
 | 
						|
#define ENOUGH_WORK_IN_SECONDS 2
 | 
						|
#define BUF_SIZE (8<<20)
 | 
						|
#define BUF_PAGES ((8<<20)>>12)
 | 
						|
drm_intel_bo *dst, *dst2;
 | 
						|
 | 
						|
/* returns time diff in milliseconds */
 | 
						|
static int64_t
 | 
						|
do_time_diff(struct timespec *end, struct timespec *start)
 | 
						|
{
 | 
						|
	int64_t ret;
 | 
						|
	ret = (MSEC_PER_SEC * difftime(end->tv_sec, start->tv_sec)) +
 | 
						|
	      ((end->tv_nsec/NSEC_PER_MSEC) - (start->tv_nsec/NSEC_PER_MSEC));
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int
 | 
						|
gem_bo_wait_timeout(int fd, uint32_t handle, int64_t *timeout_ns)
 | 
						|
{
 | 
						|
	struct drm_i915_gem_wait wait;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	igt_assert(timeout_ns);
 | 
						|
 | 
						|
	wait.bo_handle = handle;
 | 
						|
	wait.timeout_ns = *timeout_ns;
 | 
						|
	wait.flags = 0;
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
 | 
						|
	*timeout_ns = wait.timeout_ns;
 | 
						|
 | 
						|
	return ret ? -errno : 0;
 | 
						|
}
 | 
						|
 | 
						|
static void blt_color_fill(struct intel_batchbuffer *batch,
 | 
						|
			   drm_intel_bo *buf,
 | 
						|
			   const unsigned int pages)
 | 
						|
{
 | 
						|
	const unsigned short height = pages/4;
 | 
						|
	const unsigned short width =  4096;
 | 
						|
 | 
						|
	COLOR_BLIT_COPY_BATCH_START(COLOR_BLT_WRITE_ALPHA |
 | 
						|
				    XY_COLOR_BLT_WRITE_RGB);
 | 
						|
	OUT_BATCH((3 << 24)	| /* 32 Bit Color */
 | 
						|
		  (0xF0 << 16)	| /* Raster OP copy background register */
 | 
						|
		  0);		  /* Dest pitch is 0 */
 | 
						|
	OUT_BATCH(0);
 | 
						|
	OUT_BATCH(width << 16	|
 | 
						|
		  height);
 | 
						|
	OUT_RELOC_FENCED(buf, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0);
 | 
						|
	OUT_BATCH(rand()); /* random pattern */
 | 
						|
	ADVANCE_BATCH();
 | 
						|
}
 | 
						|
 | 
						|
static void render_timeout(int fd)
 | 
						|
{
 | 
						|
	drm_intel_bufmgr *bufmgr;
 | 
						|
	struct intel_batchbuffer *batch;
 | 
						|
	int64_t timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
 | 
						|
	int64_t negative_timeout = -1;
 | 
						|
	int ret;
 | 
						|
	const bool do_signals = true; /* signals will seem to make the operation
 | 
						|
				       * use less process CPU time */
 | 
						|
	bool done = false;
 | 
						|
	int i, iter = 1;
 | 
						|
 | 
						|
	igt_skip_on_simulation();
 | 
						|
 | 
						|
	bufmgr = drm_intel_bufmgr_gem_init(fd, 4096);
 | 
						|
	drm_intel_bufmgr_gem_enable_reuse(bufmgr);
 | 
						|
	batch = intel_batchbuffer_alloc(bufmgr, intel_get_drm_devid(fd));
 | 
						|
 | 
						|
	dst = drm_intel_bo_alloc(bufmgr, "dst", BUF_SIZE, 4096);
 | 
						|
	dst2 = drm_intel_bo_alloc(bufmgr, "dst2", BUF_SIZE, 4096);
 | 
						|
 | 
						|
	igt_skip_on_f(gem_bo_wait_timeout(fd, dst->handle, &timeout) == -EINVAL,
 | 
						|
		      "kernel doesn't support wait_timeout, skipping test\n");
 | 
						|
	timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
 | 
						|
 | 
						|
	/* Figure out a rough number of fills required to consume 1 second of
 | 
						|
	 * GPU work.
 | 
						|
	 */
 | 
						|
	do {
 | 
						|
		struct timespec start, end;
 | 
						|
		long diff;
 | 
						|
 | 
						|
#ifndef CLOCK_MONOTONIC_RAW
 | 
						|
#define CLOCK_MONOTONIC_RAW CLOCK_MONOTONIC
 | 
						|
#endif
 | 
						|
 | 
						|
		igt_assert(clock_gettime(CLOCK_MONOTONIC_RAW, &start) == 0);
 | 
						|
		for (i = 0; i < iter; i++)
 | 
						|
			blt_color_fill(batch, dst, BUF_PAGES);
 | 
						|
		intel_batchbuffer_flush(batch);
 | 
						|
		drm_intel_bo_wait_rendering(dst);
 | 
						|
		igt_assert(clock_gettime(CLOCK_MONOTONIC_RAW, &end) == 0);
 | 
						|
 | 
						|
		diff = do_time_diff(&end, &start);
 | 
						|
		igt_assert(diff >= 0);
 | 
						|
 | 
						|
		if ((diff / MSEC_PER_SEC) > ENOUGH_WORK_IN_SECONDS)
 | 
						|
			done = true;
 | 
						|
		else
 | 
						|
			iter <<= 1;
 | 
						|
	} while (!done && iter < 1000000);
 | 
						|
 | 
						|
	igt_assert_lt(iter, 1000000);
 | 
						|
 | 
						|
	igt_info("%d iters is enough work\n", iter);
 | 
						|
	gem_quiescent_gpu(fd);
 | 
						|
	if (do_signals)
 | 
						|
		igt_fork_signal_helper();
 | 
						|
 | 
						|
	/* We should be able to do half as much work in the same amount of time,
 | 
						|
	 * but because we might schedule almost twice as much as required, we
 | 
						|
	 * might accidentally time out. Hence add some fudge. */
 | 
						|
	for (i = 0; i < iter/3; i++)
 | 
						|
		blt_color_fill(batch, dst2, BUF_PAGES);
 | 
						|
 | 
						|
	intel_batchbuffer_flush(batch);
 | 
						|
	igt_assert(gem_bo_busy(fd, dst2->handle) == true);
 | 
						|
 | 
						|
	igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), 0);
 | 
						|
	igt_assert(gem_bo_busy(fd, dst2->handle) == false);
 | 
						|
	igt_assert_neq(timeout, 0);
 | 
						|
	if (timeout ==  (ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC))
 | 
						|
		igt_info("Buffer was already done!\n");
 | 
						|
	else {
 | 
						|
		igt_info("Finished with %" PRIu64 " time remaining\n", timeout);
 | 
						|
	}
 | 
						|
 | 
						|
	/* check that polling with timeout=0 works. */
 | 
						|
	timeout = 0;
 | 
						|
	igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), 0);
 | 
						|
	igt_assert_eq(timeout, 0);
 | 
						|
 | 
						|
	/* Now check that we correctly time out, twice the auto-tune load should
 | 
						|
	 * be good enough. */
 | 
						|
	timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
 | 
						|
	for (i = 0; i < iter*2; i++)
 | 
						|
		blt_color_fill(batch, dst2, BUF_PAGES);
 | 
						|
 | 
						|
	intel_batchbuffer_flush(batch);
 | 
						|
 | 
						|
	ret = gem_bo_wait_timeout(fd, dst2->handle, &timeout);
 | 
						|
	igt_assert_eq(ret, -ETIME);
 | 
						|
	igt_assert_eq(timeout, 0);
 | 
						|
	igt_assert(gem_bo_busy(fd, dst2->handle) == true);
 | 
						|
 | 
						|
	/* check that polling with timeout=0 works. */
 | 
						|
	timeout = 0;
 | 
						|
	igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), -ETIME);
 | 
						|
	igt_assert_eq(timeout, 0);
 | 
						|
 | 
						|
 | 
						|
	/* Now check that we can pass negative (infinite) timeouts. */
 | 
						|
	negative_timeout = -1;
 | 
						|
	for (i = 0; i < iter; i++)
 | 
						|
		blt_color_fill(batch, dst2, BUF_PAGES);
 | 
						|
 | 
						|
	intel_batchbuffer_flush(batch);
 | 
						|
 | 
						|
	igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &negative_timeout), 0);
 | 
						|
	igt_assert_eq(negative_timeout, -1); /* infinity always remains */
 | 
						|
	igt_assert(gem_bo_busy(fd, dst2->handle) == false);
 | 
						|
 | 
						|
	if (do_signals)
 | 
						|
		igt_stop_signal_helper();
 | 
						|
	drm_intel_bo_unreference(dst2);
 | 
						|
	drm_intel_bo_unreference(dst);
 | 
						|
	intel_batchbuffer_free(batch);
 | 
						|
	drm_intel_bufmgr_destroy(bufmgr);
 | 
						|
}
 | 
						|
 | 
						|
static void invalid_flags(int fd)
 | 
						|
{
 | 
						|
	struct drm_i915_gem_wait wait;
 | 
						|
	int ret;
 | 
						|
	uint32_t handle;
 | 
						|
 | 
						|
	handle = gem_create(fd, 4096);
 | 
						|
 | 
						|
	wait.bo_handle = handle;
 | 
						|
	wait.timeout_ns = 1;
 | 
						|
	/* NOTE: This test intentionally tests for just the next available flag.
 | 
						|
	 * Don't "fix" this testcase without the ABI testcases for new flags
 | 
						|
	 * first. */
 | 
						|
	wait.flags = 1;
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
 | 
						|
 | 
						|
	igt_assert(ret != 0 && errno == EINVAL);
 | 
						|
 | 
						|
	gem_close(fd, handle);
 | 
						|
}
 | 
						|
 | 
						|
static void invalid_buf(int fd)
 | 
						|
{
 | 
						|
	struct drm_i915_gem_wait wait;
 | 
						|
	int ret;
 | 
						|
 | 
						|
	wait.bo_handle = 0;
 | 
						|
	wait.timeout_ns = 1;
 | 
						|
	wait.flags = 0;
 | 
						|
	ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
 | 
						|
 | 
						|
	igt_assert(ret != 0 && errno == ENOENT);
 | 
						|
}
 | 
						|
 | 
						|
int drm_fd;
 | 
						|
 | 
						|
igt_main
 | 
						|
{
 | 
						|
	igt_fixture
 | 
						|
		drm_fd = drm_open_driver(DRIVER_INTEL);
 | 
						|
 | 
						|
	igt_subtest("render_timeout")
 | 
						|
		render_timeout(drm_fd);
 | 
						|
 | 
						|
	igt_subtest("invalid-flags")
 | 
						|
		invalid_flags(drm_fd);
 | 
						|
 | 
						|
	igt_subtest("invalid-buf")
 | 
						|
		invalid_buf(drm_fd);
 | 
						|
 | 
						|
	igt_fixture
 | 
						|
		close(drm_fd);
 | 
						|
}
 |