mirror of
https://github.com/tiagovignatti/intel-gpu-tools.git
synced 2025-06-19 22:06:13 +00:00
This patch adds test_correct_cpu_write, which maps the texture buffer through a prime fd and then writes directly to it using the CPU. It stresses the driver to guarantee cache synchronization among the different domains. This test also adds test_forked_cpu_write, which creates the GEM bo in one process and pass the prime handle of the it to another process, which in turn uses the handle only to map and write. Roughly speaking this test simulates Chrome OS architecture, where the Web content ("unpriviledged process") maps and CPU-draws a buffer, which was previously allocated in the GPU process ("priviledged process"). This requires kernel modifications (Daniel Thompson's "drm: prime: Honour O_RDWR during prime-handle-to-fd") and therefore prime_handle_to_fd_for_mmap is added to fail in case these lack. Also, upcoming tests (e.g. next patch) are going to use it as well, so make it public and available in the lib. v2: adds prime_handle_to_fd_with_mmap for skipping test in older kernels and test for invalid flags. Signed-off-by: Tiago Vignatti <tiago.vignatti@intel.com>
495 lines
12 KiB
C
495 lines
12 KiB
C
/*
|
|
* Copyright © 2014 Intel Corporation
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*
|
|
* Authors:
|
|
* Rob Bradford <rob at linux.intel.com>
|
|
* Tiago Vignatti <tiago.vignatti at intel.com>
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* Testcase: Check whether mmap()ing dma-buf works
|
|
*/
|
|
#define _GNU_SOURCE
|
|
#include <unistd.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <fcntl.h>
|
|
#include <inttypes.h>
|
|
#include <errno.h>
|
|
#include <sys/stat.h>
|
|
#include <sys/ioctl.h>
|
|
#include <pthread.h>
|
|
|
|
#include "drm.h"
|
|
#include "i915_drm.h"
|
|
#include "drmtest.h"
|
|
#include "igt_debugfs.h"
|
|
#include "ioctl_wrappers.h"
|
|
|
|
#define BO_SIZE (16*1024)
|
|
|
|
static int fd;
|
|
|
|
char pattern[] = {0xff, 0x00, 0x00, 0x00,
|
|
0x00, 0xff, 0x00, 0x00,
|
|
0x00, 0x00, 0xff, 0x00,
|
|
0x00, 0x00, 0x00, 0xff};
|
|
|
|
static void
|
|
fill_bo(uint32_t handle, size_t size)
|
|
{
|
|
off_t i;
|
|
for (i = 0; i < size; i+=sizeof(pattern))
|
|
{
|
|
gem_write(fd, handle, i, pattern, sizeof(pattern));
|
|
}
|
|
}
|
|
|
|
static void
|
|
fill_bo_cpu(char *ptr)
|
|
{
|
|
memcpy(ptr, pattern, sizeof(pattern));
|
|
}
|
|
|
|
static void
|
|
test_correct(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr1, *ptr2;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
|
|
/* Check correctness vs GEM_MMAP_GTT */
|
|
ptr1 = gem_mmap__gtt(fd, handle, BO_SIZE, PROT_READ);
|
|
ptr2 = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr1 != MAP_FAILED);
|
|
igt_assert(ptr2 != MAP_FAILED);
|
|
igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0);
|
|
|
|
/* Check pattern correctness */
|
|
igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0);
|
|
|
|
munmap(ptr1, BO_SIZE);
|
|
munmap(ptr2, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
static void
|
|
test_map_unmap(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
|
|
/* Unmap and remap */
|
|
munmap(ptr, BO_SIZE);
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
|
|
munmap(ptr, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
/* prime and then unprime and then prime again the same handle */
|
|
static void
|
|
test_reprime(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
|
|
close (dma_buf_fd);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
munmap(ptr, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
|
|
munmap(ptr, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
/* map from another process */
|
|
static void
|
|
test_forked(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
|
|
igt_fork(childno, 1) {
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
munmap(ptr, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
}
|
|
close(dma_buf_fd);
|
|
igt_waitchildren();
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
/* test simple CPU write */
|
|
static void
|
|
test_correct_cpu_write(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle);
|
|
|
|
/* Skip if DRM_RDWR is not supported */
|
|
igt_skip_on(errno == EINVAL);
|
|
|
|
/* Check correctness of map using write protection (PROT_WRITE) */
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
|
|
/* Fill bo using CPU */
|
|
fill_bo_cpu(ptr);
|
|
|
|
/* Check pattern correctness */
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
|
|
munmap(ptr, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
/* map from another process and then write using CPU */
|
|
static void
|
|
test_forked_cpu_write(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle);
|
|
|
|
/* Skip if DRM_RDWR is not supported */
|
|
igt_skip_on(errno == EINVAL);
|
|
|
|
igt_fork(childno, 1) {
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE , MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
fill_bo_cpu(ptr);
|
|
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
munmap(ptr, BO_SIZE);
|
|
close(dma_buf_fd);
|
|
}
|
|
close(dma_buf_fd);
|
|
igt_waitchildren();
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
static void
|
|
test_refcounting(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
/* Close gem object before mapping */
|
|
gem_close(fd, handle);
|
|
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
munmap(ptr, BO_SIZE);
|
|
close (dma_buf_fd);
|
|
}
|
|
|
|
/* dup before mmap */
|
|
static void
|
|
test_dup(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
|
|
dma_buf_fd = dup(prime_handle_to_fd(fd, handle));
|
|
igt_assert(errno == 0);
|
|
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr != MAP_FAILED);
|
|
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
|
|
munmap(ptr, BO_SIZE);
|
|
gem_close(fd, handle);
|
|
close (dma_buf_fd);
|
|
}
|
|
|
|
/* Used for error case testing to avoid wrapper */
|
|
static int prime_handle_to_fd_no_assert(uint32_t handle, int flags, int *fd_out)
|
|
{
|
|
struct drm_prime_handle args;
|
|
int ret;
|
|
|
|
args.handle = handle;
|
|
args.flags = flags;
|
|
args.fd = -1;
|
|
|
|
ret = drmIoctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args);
|
|
if (ret)
|
|
ret = errno;
|
|
*fd_out = args.fd;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* test for mmap(dma_buf_export(userptr)) */
|
|
static void
|
|
test_userptr(void)
|
|
{
|
|
int ret, dma_buf_fd;
|
|
void *ptr;
|
|
uint32_t handle;
|
|
|
|
/* create userptr bo */
|
|
ret = posix_memalign(&ptr, 4096, BO_SIZE);
|
|
igt_assert_eq(ret, 0);
|
|
|
|
/* we are not allowed to export unsynchronized userptr. Just create a normal
|
|
* one */
|
|
gem_userptr(fd, (uint32_t *)ptr, BO_SIZE, 0, 0, &handle);
|
|
|
|
/* export userptr */
|
|
ret = prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd);
|
|
if (ret) {
|
|
igt_assert(ret == EINVAL || ret == ENODEV);
|
|
goto free_userptr;
|
|
} else {
|
|
igt_assert_eq(ret, 0);
|
|
igt_assert_lte(0, dma_buf_fd);
|
|
}
|
|
|
|
/* a userptr doesn't have the obj->base.filp, but can be exported via
|
|
* dma-buf, so make sure it fails here */
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr == MAP_FAILED && errno == ENODEV);
|
|
free_userptr:
|
|
gem_close(fd, handle);
|
|
close(dma_buf_fd);
|
|
}
|
|
|
|
static void
|
|
test_errors(void)
|
|
{
|
|
int i, dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
int invalid_flags[] = {DRM_CLOEXEC - 1, DRM_CLOEXEC + 1,
|
|
DRM_RDWR - 1, DRM_RDWR + 1};
|
|
|
|
/* Test for invalid flags */
|
|
handle = gem_create(fd, BO_SIZE);
|
|
for (i = 0; i < sizeof(invalid_flags) / sizeof(invalid_flags[0]); i++) {
|
|
prime_handle_to_fd_no_assert(handle, invalid_flags[i], &dma_buf_fd);
|
|
igt_assert_eq(errno, EINVAL);
|
|
errno = 0;
|
|
}
|
|
|
|
/* Close gem object before priming */
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
gem_close(fd, handle);
|
|
prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd);
|
|
igt_assert(dma_buf_fd == -1 && errno == ENOENT);
|
|
errno = 0;
|
|
|
|
/* close fd before mapping */
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
close(dma_buf_fd);
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr == MAP_FAILED && errno == EBADF);
|
|
errno = 0;
|
|
gem_close(fd, handle);
|
|
|
|
/* Map too big */
|
|
handle = gem_create(fd, BO_SIZE);
|
|
fill_bo(handle, BO_SIZE);
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
ptr = mmap(NULL, BO_SIZE * 2, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
igt_assert(ptr == MAP_FAILED && errno == EINVAL);
|
|
errno = 0;
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
|
|
/* Overlapping the end of the buffer */
|
|
handle = gem_create(fd, BO_SIZE);
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
igt_assert(errno == 0);
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, BO_SIZE / 2);
|
|
igt_assert(ptr == MAP_FAILED && errno == EINVAL);
|
|
errno = 0;
|
|
close(dma_buf_fd);
|
|
gem_close(fd, handle);
|
|
}
|
|
|
|
static void
|
|
test_aperture_limit(void)
|
|
{
|
|
int dma_buf_fd1, dma_buf_fd2;
|
|
char *ptr1, *ptr2;
|
|
uint32_t handle1, handle2;
|
|
/* Two buffers the sum of which > mappable aperture */
|
|
uint64_t size1 = (gem_mappable_aperture_size() * 7) / 8;
|
|
uint64_t size2 = (gem_mappable_aperture_size() * 3) / 8;
|
|
|
|
handle1 = gem_create(fd, size1);
|
|
fill_bo(handle1, BO_SIZE);
|
|
|
|
dma_buf_fd1 = prime_handle_to_fd(fd, handle1);
|
|
igt_assert(errno == 0);
|
|
ptr1 = mmap(NULL, size1, PROT_READ, MAP_SHARED, dma_buf_fd1, 0);
|
|
igt_assert(ptr1 != MAP_FAILED);
|
|
igt_assert(memcmp(ptr1, pattern, sizeof(pattern)) == 0);
|
|
|
|
handle2 = gem_create(fd, size1);
|
|
fill_bo(handle2, BO_SIZE);
|
|
dma_buf_fd2 = prime_handle_to_fd(fd, handle2);
|
|
igt_assert(errno == 0);
|
|
ptr2 = mmap(NULL, size2, PROT_READ, MAP_SHARED, dma_buf_fd2, 0);
|
|
igt_assert(ptr2 != MAP_FAILED);
|
|
igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0);
|
|
|
|
igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0);
|
|
|
|
munmap(ptr1, size1);
|
|
munmap(ptr2, size2);
|
|
close(dma_buf_fd1);
|
|
close(dma_buf_fd2);
|
|
gem_close(fd, handle1);
|
|
gem_close(fd, handle2);
|
|
}
|
|
|
|
static int
|
|
check_for_dma_buf_mmap(void)
|
|
{
|
|
int dma_buf_fd;
|
|
char *ptr;
|
|
uint32_t handle;
|
|
int ret = 1;
|
|
|
|
handle = gem_create(fd, BO_SIZE);
|
|
dma_buf_fd = prime_handle_to_fd(fd, handle);
|
|
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
|
|
if (ptr != MAP_FAILED)
|
|
ret = 0;
|
|
munmap(ptr, BO_SIZE);
|
|
gem_close(fd, handle);
|
|
close(dma_buf_fd);
|
|
return ret;
|
|
}
|
|
|
|
igt_main
|
|
{
|
|
struct {
|
|
const char *name;
|
|
void (*fn)(void);
|
|
} tests[] = {
|
|
{ "test_correct", test_correct },
|
|
{ "test_map_unmap", test_map_unmap },
|
|
{ "test_reprime", test_reprime },
|
|
{ "test_forked", test_forked },
|
|
{ "test_correct_cpu_write", test_correct_cpu_write },
|
|
{ "test_forked_cpu_write", test_forked_cpu_write },
|
|
{ "test_refcounting", test_refcounting },
|
|
{ "test_dup", test_dup },
|
|
{ "test_userptr", test_userptr },
|
|
{ "test_errors", test_errors },
|
|
{ "test_aperture_limit", test_aperture_limit },
|
|
};
|
|
int i;
|
|
|
|
igt_fixture {
|
|
fd = drm_open_driver(DRIVER_INTEL);
|
|
errno = 0;
|
|
}
|
|
|
|
igt_skip_on((check_for_dma_buf_mmap() != 0));
|
|
|
|
for (i = 0; i < ARRAY_SIZE(tests); i++) {
|
|
igt_subtest(tests[i].name)
|
|
tests[i].fn();
|
|
}
|
|
|
|
igt_fixture
|
|
close(fd);
|
|
}
|