blob: 67a6a232ad2a5deb7d8983bb421ed522fdc10d09 [file] [log] [blame]
/*
* Copyright © 2014 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*
* Authors:
* Rob Bradford <rob at linux.intel.com>
* Tiago Vignatti <tiago.vignatti at intel.com>
*
*/
/*
* Testcase: Check whether mmap()ing dma-buf works
*/
#include <unistd.h>
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <fcntl.h>
#include <inttypes.h>
#include <errno.h>
#include <sys/stat.h>
#include <sys/ioctl.h>
#include <pthread.h>
#include "drm.h"
#include "i915_drm.h"
#include "drmtest.h"
#include "igt_debugfs.h"
#include "ioctl_wrappers.h"
#define BO_SIZE (16*1024)
static int fd;
char pattern[] = {0xff, 0x00, 0x00, 0x00,
0x00, 0xff, 0x00, 0x00,
0x00, 0x00, 0xff, 0x00,
0x00, 0x00, 0x00, 0xff};
static void
fill_bo(uint32_t handle, size_t size)
{
off_t i;
for (i = 0; i < size; i+=sizeof(pattern))
{
gem_write(fd, handle, i, pattern, sizeof(pattern));
}
}
static void
fill_bo_cpu(char *ptr)
{
memcpy(ptr, pattern, sizeof(pattern));
}
static void
test_correct(void)
{
int dma_buf_fd;
char *ptr1, *ptr2;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
/* Check correctness vs GEM_MMAP_GTT */
ptr1 = gem_mmap__gtt(fd, handle, BO_SIZE, PROT_READ);
ptr2 = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr1 != MAP_FAILED);
igt_assert(ptr2 != MAP_FAILED);
igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0);
/* Check pattern correctness */
igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0);
munmap(ptr1, BO_SIZE);
munmap(ptr2, BO_SIZE);
close(dma_buf_fd);
gem_close(fd, handle);
}
static void
test_map_unmap(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
/* Unmap and remap */
munmap(ptr, BO_SIZE);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close(dma_buf_fd);
gem_close(fd, handle);
}
/* prime and then unprime and then prime again the same handle */
static void
test_reprime(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
close (dma_buf_fd);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close(dma_buf_fd);
gem_close(fd, handle);
}
/* map from another process */
static void
test_forked(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
igt_fork(childno, 1) {
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close(dma_buf_fd);
}
close(dma_buf_fd);
igt_waitchildren();
gem_close(fd, handle);
}
/* test simple CPU write */
static void
test_correct_cpu_write(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle);
/* Skip if DRM_RDWR is not supported */
igt_skip_on(errno == EINVAL);
/* Check correctness of map using write protection (PROT_WRITE) */
ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
/* Fill bo using CPU */
fill_bo_cpu(ptr);
/* Check pattern correctness */
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close(dma_buf_fd);
gem_close(fd, handle);
}
/* map from another process and then write using CPU */
static void
test_forked_cpu_write(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
dma_buf_fd = prime_handle_to_fd_for_mmap(fd, handle);
/* Skip if DRM_RDWR is not supported */
igt_skip_on(errno == EINVAL);
igt_fork(childno, 1) {
ptr = mmap(NULL, BO_SIZE, PROT_READ | PROT_WRITE , MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
fill_bo_cpu(ptr);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close(dma_buf_fd);
}
close(dma_buf_fd);
igt_waitchildren();
gem_close(fd, handle);
}
static void
test_refcounting(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
/* Close gem object before mapping */
gem_close(fd, handle);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
close (dma_buf_fd);
}
/* dup before mmap */
static void
test_dup(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = dup(prime_handle_to_fd(fd, handle));
igt_assert(errno == 0);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr != MAP_FAILED);
igt_assert(memcmp(ptr, pattern, sizeof(pattern)) == 0);
munmap(ptr, BO_SIZE);
gem_close(fd, handle);
close (dma_buf_fd);
}
/* Used for error case testing to avoid wrapper */
static int prime_handle_to_fd_no_assert(uint32_t handle, int flags, int *fd_out)
{
struct drm_prime_handle args;
int ret;
args.handle = handle;
args.flags = flags;
args.fd = -1;
ret = drmIoctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args);
if (ret)
ret = errno;
*fd_out = args.fd;
return ret;
}
static bool has_userptr(void)
{
uint32_t handle = 0;
void *ptr;
igt_assert(posix_memalign(&ptr, 4096, 4096) == 0);
if ( __gem_userptr(fd, ptr, 4096, 0, 0, &handle) == 0)
gem_close(fd, handle);
free(ptr);
return handle;
}
/* test for mmap(dma_buf_export(userptr)) */
static void
test_userptr(void)
{
int ret, dma_buf_fd;
void *ptr;
uint32_t handle;
igt_require(has_userptr());
/* create userptr bo */
ret = posix_memalign(&ptr, 4096, BO_SIZE);
igt_assert_eq(ret, 0);
/* we are not allowed to export unsynchronized userptr. Just create a normal
* one */
gem_userptr(fd, (uint32_t *)ptr, BO_SIZE, 0, 0, &handle);
/* export userptr */
ret = prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd);
if (ret) {
igt_assert(ret == EINVAL || ret == ENODEV);
goto free_userptr;
} else {
igt_assert_eq(ret, 0);
igt_assert_lte(0, dma_buf_fd);
}
/* a userptr doesn't have the obj->base.filp, but can be exported via
* dma-buf, so make sure it fails here */
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr == MAP_FAILED && errno == ENODEV);
free_userptr:
gem_close(fd, handle);
close(dma_buf_fd);
}
static void
test_errors(void)
{
int i, dma_buf_fd;
char *ptr;
uint32_t handle;
int invalid_flags[] = {DRM_CLOEXEC - 1, DRM_CLOEXEC + 1,
DRM_RDWR - 1, DRM_RDWR + 1};
/* Test for invalid flags */
handle = gem_create(fd, BO_SIZE);
for (i = 0; i < sizeof(invalid_flags) / sizeof(invalid_flags[0]); i++) {
prime_handle_to_fd_no_assert(handle, invalid_flags[i], &dma_buf_fd);
igt_assert_eq(errno, EINVAL);
errno = 0;
}
/* Close gem object before priming */
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
gem_close(fd, handle);
prime_handle_to_fd_no_assert(handle, DRM_CLOEXEC, &dma_buf_fd);
igt_assert(dma_buf_fd == -1 && errno == ENOENT);
errno = 0;
/* close fd before mapping */
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
close(dma_buf_fd);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr == MAP_FAILED && errno == EBADF);
errno = 0;
gem_close(fd, handle);
/* Map too big */
handle = gem_create(fd, BO_SIZE);
fill_bo(handle, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
ptr = mmap(NULL, BO_SIZE * 2, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
igt_assert(ptr == MAP_FAILED && errno == EINVAL);
errno = 0;
close(dma_buf_fd);
gem_close(fd, handle);
/* Overlapping the end of the buffer */
handle = gem_create(fd, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
igt_assert(errno == 0);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, BO_SIZE / 2);
igt_assert(ptr == MAP_FAILED && errno == EINVAL);
errno = 0;
close(dma_buf_fd);
gem_close(fd, handle);
}
/* Test for invalid flags on sync ioctl */
static void
test_invalid_sync_flags(void)
{
int i, dma_buf_fd;
uint32_t handle;
struct local_dma_buf_sync sync;
int invalid_flags[] = {-1,
0x00,
LOCAL_DMA_BUF_SYNC_RW + 1,
LOCAL_DMA_BUF_SYNC_VALID_FLAGS_MASK + 1};
handle = gem_create(fd, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
for (i = 0; i < sizeof(invalid_flags) / sizeof(invalid_flags[0]); i++) {
memset(&sync, 0, sizeof(sync));
sync.flags = invalid_flags[i];
drmIoctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync);
igt_assert_eq(errno, EINVAL);
errno = 0;
}
}
static void
test_aperture_limit(void)
{
int dma_buf_fd1, dma_buf_fd2;
char *ptr1, *ptr2;
uint32_t handle1, handle2;
/* Two buffers the sum of which > mappable aperture */
uint64_t size1 = (gem_mappable_aperture_size() * 7) / 8;
uint64_t size2 = (gem_mappable_aperture_size() * 3) / 8;
handle1 = gem_create(fd, size1);
fill_bo(handle1, BO_SIZE);
dma_buf_fd1 = prime_handle_to_fd(fd, handle1);
igt_assert(errno == 0);
ptr1 = mmap(NULL, size1, PROT_READ, MAP_SHARED, dma_buf_fd1, 0);
igt_assert(ptr1 != MAP_FAILED);
igt_assert(memcmp(ptr1, pattern, sizeof(pattern)) == 0);
handle2 = gem_create(fd, size1);
fill_bo(handle2, BO_SIZE);
dma_buf_fd2 = prime_handle_to_fd(fd, handle2);
igt_assert(errno == 0);
ptr2 = mmap(NULL, size2, PROT_READ, MAP_SHARED, dma_buf_fd2, 0);
igt_assert(ptr2 != MAP_FAILED);
igt_assert(memcmp(ptr2, pattern, sizeof(pattern)) == 0);
igt_assert(memcmp(ptr1, ptr2, BO_SIZE) == 0);
munmap(ptr1, size1);
munmap(ptr2, size2);
close(dma_buf_fd1);
close(dma_buf_fd2);
gem_close(fd, handle1);
gem_close(fd, handle2);
}
static int
check_for_dma_buf_mmap(void)
{
int dma_buf_fd;
char *ptr;
uint32_t handle;
int ret = 1;
handle = gem_create(fd, BO_SIZE);
dma_buf_fd = prime_handle_to_fd(fd, handle);
ptr = mmap(NULL, BO_SIZE, PROT_READ, MAP_SHARED, dma_buf_fd, 0);
if (ptr != MAP_FAILED)
ret = 0;
munmap(ptr, BO_SIZE);
gem_close(fd, handle);
close(dma_buf_fd);
return ret;
}
igt_main
{
struct {
const char *name;
void (*fn)(void);
} tests[] = {
{ "test_correct", test_correct },
{ "test_map_unmap", test_map_unmap },
{ "test_reprime", test_reprime },
{ "test_forked", test_forked },
{ "test_correct_cpu_write", test_correct_cpu_write },
{ "test_forked_cpu_write", test_forked_cpu_write },
{ "test_refcounting", test_refcounting },
{ "test_dup", test_dup },
{ "test_userptr", test_userptr },
{ "test_errors", test_errors },
{ "test_invalid_sync_flags", test_invalid_sync_flags },
{ "test_aperture_limit", test_aperture_limit },
};
int i;
igt_fixture {
fd = drm_open_driver(DRIVER_INTEL);
igt_skip_on((check_for_dma_buf_mmap() != 0));
errno = 0;
}
for (i = 0; i < ARRAY_SIZE(tests); i++) {
igt_subtest(tests[i].name)
tests[i].fn();
}
igt_fixture
close(fd);
}