mirror of
https://github.com/tiagovignatti/intel-gpu-tools.git
synced 2025-06-07 16:06:25 +00:00
Apply the new API to all call sites within the test suite using the following semantic patch: // Semantic patch for replacing drm_open_any* with arch-specific drm_open_driver* calls @@ identifier i =~ "\bdrm_open_any\b"; @@ - i() + drm_open_driver(DRIVER_INTEL) @@ identifier i =~ "\bdrm_open_any_master\b"; @@ - i() + drm_open_driver_master(DRIVER_INTEL) @@ identifier i =~ "\bdrm_open_any_render\b"; @@ - i() + drm_open_driver_render(DRIVER_INTEL) @@ identifier i =~ "\b__drm_open_any\b"; @@ - i() + __drm_open_driver(DRIVER_INTEL) Signed-off-by: Micah Fedke <micah.fedke@collabora.co.uk> Signed-off-by: Thomas Wood <thomas.wood@intel.com>
277 lines
7.5 KiB
C
277 lines
7.5 KiB
C
/*
|
|
* Copyright © 2012 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Ben Widawsky <ben@bwidawsk.net>
|
|
*
|
|
*/
|
|
|
|
#include "igt.h"
|
|
#include <stdio.h>
|
|
#include <time.h>
|
|
#include <stdlib.h>
|
|
#include <sys/ioctl.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <fcntl.h>
|
|
#include <inttypes.h>
|
|
#include <errno.h>
|
|
#include <sys/stat.h>
|
|
#include <sys/time.h>
|
|
|
|
#include <drm.h>
|
|
|
|
#include "intel_bufmgr.h"
|
|
|
|
#define MSEC_PER_SEC 1000L
|
|
#define USEC_PER_MSEC 1000L
|
|
#define NSEC_PER_USEC 1000L
|
|
#define NSEC_PER_MSEC 1000000L
|
|
#define USEC_PER_SEC 1000000L
|
|
#define NSEC_PER_SEC 1000000000L
|
|
|
|
#define ENOUGH_WORK_IN_SECONDS 2
|
|
#define BUF_SIZE (8<<20)
|
|
#define BUF_PAGES ((8<<20)>>12)
|
|
drm_intel_bo *dst, *dst2;
|
|
|
|
/* returns time diff in milliseconds */
|
|
static int64_t
|
|
do_time_diff(struct timespec *end, struct timespec *start)
|
|
{
|
|
int64_t ret;
|
|
ret = (MSEC_PER_SEC * difftime(end->tv_sec, start->tv_sec)) +
|
|
((end->tv_nsec/NSEC_PER_MSEC) - (start->tv_nsec/NSEC_PER_MSEC));
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
gem_bo_wait_timeout(int fd, uint32_t handle, int64_t *timeout_ns)
|
|
{
|
|
struct drm_i915_gem_wait wait;
|
|
int ret;
|
|
|
|
igt_assert(timeout_ns);
|
|
|
|
wait.bo_handle = handle;
|
|
wait.timeout_ns = *timeout_ns;
|
|
wait.flags = 0;
|
|
ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
|
|
*timeout_ns = wait.timeout_ns;
|
|
|
|
return ret ? -errno : 0;
|
|
}
|
|
|
|
static void blt_color_fill(struct intel_batchbuffer *batch,
|
|
drm_intel_bo *buf,
|
|
const unsigned int pages)
|
|
{
|
|
const unsigned short height = pages/4;
|
|
const unsigned short width = 4096;
|
|
|
|
COLOR_BLIT_COPY_BATCH_START(COLOR_BLT_WRITE_ALPHA |
|
|
XY_COLOR_BLT_WRITE_RGB);
|
|
OUT_BATCH((3 << 24) | /* 32 Bit Color */
|
|
(0xF0 << 16) | /* Raster OP copy background register */
|
|
0); /* Dest pitch is 0 */
|
|
OUT_BATCH(0);
|
|
OUT_BATCH(width << 16 |
|
|
height);
|
|
OUT_RELOC_FENCED(buf, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0);
|
|
OUT_BATCH(rand()); /* random pattern */
|
|
ADVANCE_BATCH();
|
|
}
|
|
|
|
static void render_timeout(int fd)
|
|
{
|
|
drm_intel_bufmgr *bufmgr;
|
|
struct intel_batchbuffer *batch;
|
|
int64_t timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
|
|
int64_t negative_timeout = -1;
|
|
int ret;
|
|
const bool do_signals = true; /* signals will seem to make the operation
|
|
* use less process CPU time */
|
|
bool done = false;
|
|
int i, iter = 1;
|
|
|
|
igt_skip_on_simulation();
|
|
|
|
bufmgr = drm_intel_bufmgr_gem_init(fd, 4096);
|
|
drm_intel_bufmgr_gem_enable_reuse(bufmgr);
|
|
batch = intel_batchbuffer_alloc(bufmgr, intel_get_drm_devid(fd));
|
|
|
|
dst = drm_intel_bo_alloc(bufmgr, "dst", BUF_SIZE, 4096);
|
|
dst2 = drm_intel_bo_alloc(bufmgr, "dst2", BUF_SIZE, 4096);
|
|
|
|
igt_skip_on_f(gem_bo_wait_timeout(fd, dst->handle, &timeout) == -EINVAL,
|
|
"kernel doesn't support wait_timeout, skipping test\n");
|
|
timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
|
|
|
|
/* Figure out a rough number of fills required to consume 1 second of
|
|
* GPU work.
|
|
*/
|
|
do {
|
|
struct timespec start, end;
|
|
long diff;
|
|
|
|
#ifndef CLOCK_MONOTONIC_RAW
|
|
#define CLOCK_MONOTONIC_RAW CLOCK_MONOTONIC
|
|
#endif
|
|
|
|
igt_assert(clock_gettime(CLOCK_MONOTONIC_RAW, &start) == 0);
|
|
for (i = 0; i < iter; i++)
|
|
blt_color_fill(batch, dst, BUF_PAGES);
|
|
intel_batchbuffer_flush(batch);
|
|
drm_intel_bo_wait_rendering(dst);
|
|
igt_assert(clock_gettime(CLOCK_MONOTONIC_RAW, &end) == 0);
|
|
|
|
diff = do_time_diff(&end, &start);
|
|
igt_assert(diff >= 0);
|
|
|
|
if ((diff / MSEC_PER_SEC) > ENOUGH_WORK_IN_SECONDS)
|
|
done = true;
|
|
else
|
|
iter <<= 1;
|
|
} while (!done && iter < 1000000);
|
|
|
|
igt_assert_lt(iter, 1000000);
|
|
|
|
igt_info("%d iters is enough work\n", iter);
|
|
gem_quiescent_gpu(fd);
|
|
if (do_signals)
|
|
igt_fork_signal_helper();
|
|
|
|
/* We should be able to do half as much work in the same amount of time,
|
|
* but because we might schedule almost twice as much as required, we
|
|
* might accidentally time out. Hence add some fudge. */
|
|
for (i = 0; i < iter/3; i++)
|
|
blt_color_fill(batch, dst2, BUF_PAGES);
|
|
|
|
intel_batchbuffer_flush(batch);
|
|
igt_assert(gem_bo_busy(fd, dst2->handle) == true);
|
|
|
|
igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), 0);
|
|
igt_assert(gem_bo_busy(fd, dst2->handle) == false);
|
|
igt_assert_neq(timeout, 0);
|
|
if (timeout == (ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC))
|
|
igt_info("Buffer was already done!\n");
|
|
else {
|
|
igt_info("Finished with %" PRIu64 " time remaining\n", timeout);
|
|
}
|
|
|
|
/* check that polling with timeout=0 works. */
|
|
timeout = 0;
|
|
igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), 0);
|
|
igt_assert_eq(timeout, 0);
|
|
|
|
/* Now check that we correctly time out, twice the auto-tune load should
|
|
* be good enough. */
|
|
timeout = ENOUGH_WORK_IN_SECONDS * NSEC_PER_SEC;
|
|
for (i = 0; i < iter*2; i++)
|
|
blt_color_fill(batch, dst2, BUF_PAGES);
|
|
|
|
intel_batchbuffer_flush(batch);
|
|
|
|
ret = gem_bo_wait_timeout(fd, dst2->handle, &timeout);
|
|
igt_assert_eq(ret, -ETIME);
|
|
igt_assert_eq(timeout, 0);
|
|
igt_assert(gem_bo_busy(fd, dst2->handle) == true);
|
|
|
|
/* check that polling with timeout=0 works. */
|
|
timeout = 0;
|
|
igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &timeout), -ETIME);
|
|
igt_assert_eq(timeout, 0);
|
|
|
|
|
|
/* Now check that we can pass negative (infinite) timeouts. */
|
|
negative_timeout = -1;
|
|
for (i = 0; i < iter; i++)
|
|
blt_color_fill(batch, dst2, BUF_PAGES);
|
|
|
|
intel_batchbuffer_flush(batch);
|
|
|
|
igt_assert_eq(gem_bo_wait_timeout(fd, dst2->handle, &negative_timeout), 0);
|
|
igt_assert_eq(negative_timeout, -1); /* infinity always remains */
|
|
igt_assert(gem_bo_busy(fd, dst2->handle) == false);
|
|
|
|
if (do_signals)
|
|
igt_stop_signal_helper();
|
|
drm_intel_bo_unreference(dst2);
|
|
drm_intel_bo_unreference(dst);
|
|
intel_batchbuffer_free(batch);
|
|
drm_intel_bufmgr_destroy(bufmgr);
|
|
}
|
|
|
|
static void invalid_flags(int fd)
|
|
{
|
|
struct drm_i915_gem_wait wait;
|
|
int ret;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, 4096);
|
|
|
|
wait.bo_handle = handle;
|
|
wait.timeout_ns = 1;
|
|
/* NOTE: This test intentionally tests for just the next available flag.
|
|
* Don't "fix" this testcase without the ABI testcases for new flags
|
|
* first. */
|
|
wait.flags = 1;
|
|
ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
|
|
|
|
igt_assert(ret != 0 && errno == EINVAL);
|
|
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
static void invalid_buf(int fd)
|
|
{
|
|
struct drm_i915_gem_wait wait;
|
|
int ret;
|
|
|
|
wait.bo_handle = 0;
|
|
wait.timeout_ns = 1;
|
|
wait.flags = 0;
|
|
ret = drmIoctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait);
|
|
|
|
igt_assert(ret != 0 && errno == ENOENT);
|
|
}
|
|
|
|
int drm_fd;
|
|
|
|
igt_main
|
|
{
|
|
igt_fixture
|
|
drm_fd = drm_open_driver(DRIVER_INTEL);
|
|
|
|
igt_subtest("render_timeout")
|
|
render_timeout(drm_fd);
|
|
|
|
igt_subtest("invalid-flags")
|
|
invalid_flags(drm_fd);
|
|
|
|
igt_subtest("invalid-buf")
|
|
invalid_buf(drm_fd);
|
|
|
|
igt_fixture
|
|
close(drm_fd);
|
|
}
|