blob: 9d67a19c12ba6605fb94bc2e8496e0434081a535 [file] [log] [blame]
/*
* Copyright © 2007, 2011, 2013, 2014 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*
* Authors:
* Eric Anholt <eric@anholt.net>
* Daniel Vetter <daniel.vetter@ffwll.ch>
*
*/
#ifndef ANDROID
#define _GNU_SOURCE
#else
#include <libgen.h>
#endif
#include <stdio.h>
#include <fcntl.h>
#include <sys/stat.h>
#include <sys/ioctl.h>
#include <string.h>
#include <sys/mman.h>
#include <signal.h>
#include <pciaccess.h>
#include <getopt.h>
#include <stdlib.h>
#include <unistd.h>
#include <sys/wait.h>
#include <sys/types.h>
#include <sys/syscall.h>
#include <sys/utsname.h>
#include <termios.h>
#include <errno.h>
#include "drmtest.h"
#include "i915_drm.h"
#include "intel_chipset.h"
#include "intel_io.h"
#include "igt_debugfs.h"
#include "config.h"
#include "ioctl_wrappers.h"
/**
* SECTION:ioctl_wrappers
* @short_description: ioctl wrappers and related functions
* @title: ioctl wrappers
* @include: igt.h
*
* This helper library contains simple functions to wrap the raw drm/i915 kernel
* ioctls. The normal versions never pass any error codes to the caller and use
* igt_assert() to check for error conditions instead. For some ioctls raw
* wrappers which do pass on error codes are available. These raw wrappers have
* a __ prefix.
*
* For wrappers which check for feature bits there can also be two versions: The
* normal one simply returns a boolean to the caller. But when skipping the
* testcase entirely is the right action then it's better to use igt_skip()
* directly in the wrapper. Such functions have _require_ in their name to
* distinguish them.
*/
int (*igt_ioctl)(int fd, unsigned long request, void *arg) = drmIoctl;
/**
* gem_handle_to_libdrm_bo:
* @bufmgr: libdrm buffer manager instance
* @fd: open i915 drm file descriptor
* @name: buffer name in libdrm
* @handle: gem buffer object handle
*
* This helper function imports a raw gem buffer handle into the libdrm buffer
* manager.
*
* Returns: The imported libdrm buffer manager object.
*/
drm_intel_bo *
gem_handle_to_libdrm_bo(drm_intel_bufmgr *bufmgr, int fd, const char *name, uint32_t handle)
{
struct drm_gem_flink flink;
int ret;
drm_intel_bo *bo;
memset(&flink, 0, sizeof(handle));
flink.handle = handle;
ret = ioctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
igt_assert(ret == 0);
errno = 0;
bo = drm_intel_bo_gem_create_from_name(bufmgr, name, flink.name);
igt_assert(bo);
return bo;
}
static int
__gem_get_tiling(int fd, struct drm_i915_gem_get_tiling *arg)
{
int err;
err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_GET_TILING, arg))
err = -errno;
errno = 0;
return err;
}
/**
* gem_get_tiling:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @tiling: (out) tiling mode of the gem buffer
* @swizzle: (out) bit 6 swizzle mode
*
* This wraps the GET_TILING ioctl.
*
* Returns whether the actual physical tiling matches the reported tiling.
*/
bool
gem_get_tiling(int fd, uint32_t handle, uint32_t *tiling, uint32_t *swizzle)
{
struct drm_i915_gem_get_tiling get_tiling;
memset(&get_tiling, 0, sizeof(get_tiling));
get_tiling.handle = handle;
igt_assert_eq(__gem_get_tiling(fd, &get_tiling), 0);
*tiling = get_tiling.tiling_mode;
*swizzle = get_tiling.swizzle_mode;
return get_tiling.phys_swizzle_mode == get_tiling.swizzle_mode;
}
int __gem_set_tiling(int fd, uint32_t handle, uint32_t tiling, uint32_t stride)
{
struct drm_i915_gem_set_tiling st;
int ret;
igt_require_intel(fd);
/* The kernel doesn't know about these tiling modes, expects NONE */
if (tiling == I915_TILING_Yf || tiling == I915_TILING_Ys)
tiling = I915_TILING_NONE;
memset(&st, 0, sizeof(st));
do {
st.handle = handle;
st.tiling_mode = tiling;
st.stride = tiling ? stride : 0;
ret = ioctl(fd, DRM_IOCTL_I915_GEM_SET_TILING, &st);
} while (ret == -1 && (errno == EINTR || errno == EAGAIN));
if (ret != 0)
return -errno;
errno = 0;
igt_assert(st.tiling_mode == tiling);
return 0;
}
/**
* gem_set_tiling:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @tiling: tiling mode bits
* @stride: stride of the buffer when using a tiled mode, otherwise must be 0
*
* This wraps the SET_TILING ioctl.
*/
void gem_set_tiling(int fd, uint32_t handle, uint32_t tiling, uint32_t stride)
{
igt_assert(__gem_set_tiling(fd, handle, tiling, stride) == 0);
}
struct local_drm_i915_gem_caching {
uint32_t handle;
uint32_t caching;
};
#define LOCAL_DRM_I915_GEM_SET_CACHEING 0x2f
#define LOCAL_DRM_I915_GEM_GET_CACHEING 0x30
#define LOCAL_DRM_IOCTL_I915_GEM_SET_CACHEING \
DRM_IOW(DRM_COMMAND_BASE + LOCAL_DRM_I915_GEM_SET_CACHEING, struct local_drm_i915_gem_caching)
#define LOCAL_DRM_IOCTL_I915_GEM_GET_CACHEING \
DRM_IOWR(DRM_COMMAND_BASE + LOCAL_DRM_I915_GEM_GET_CACHEING, struct local_drm_i915_gem_caching)
static int __gem_set_caching(int fd, uint32_t handle, uint32_t caching)
{
struct local_drm_i915_gem_caching arg;
int err;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.caching = caching;
err = 0;
if (igt_ioctl(fd, LOCAL_DRM_IOCTL_I915_GEM_SET_CACHEING, &arg)) {
err = -errno;
igt_assert(errno == ENOTTY || errno == EINVAL);
}
return err;
}
/**
* gem_set_caching:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @caching: caching mode bits
*
* This wraps the SET_CACHING ioctl. Note that this function internally calls
* igt_require() when SET_CACHING isn't available, hence automatically skips the
* test. Therefore always extract test logic which uses this into its own
* subtest.
*/
void gem_set_caching(int fd, uint32_t handle, uint32_t caching)
{
igt_require(__gem_set_caching(fd, handle, caching) == 0);
errno = 0;
}
/**
* gem_get_caching:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the GET_CACHING ioctl.
*
* Returns: The current caching mode bits.
*/
uint32_t gem_get_caching(int fd, uint32_t handle)
{
struct local_drm_i915_gem_caching arg;
int ret;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
ret = ioctl(fd, LOCAL_DRM_IOCTL_I915_GEM_GET_CACHEING, &arg);
igt_assert(ret == 0);
errno = 0;
return arg.caching;
}
/**
* gem_open:
* @fd: open i915 drm file descriptor
* @name: flink buffer name
*
* This wraps the GEM_OPEN ioctl, which is used to import an flink name.
*
* Returns: gem file-private buffer handle of the open object.
*/
uint32_t gem_open(int fd, uint32_t name)
{
struct drm_gem_open open_struct;
int ret;
memset(&open_struct, 0, sizeof(open_struct));
open_struct.name = name;
ret = ioctl(fd, DRM_IOCTL_GEM_OPEN, &open_struct);
igt_assert(ret == 0);
igt_assert(open_struct.handle != 0);
errno = 0;
return open_struct.handle;
}
/**
* gem_flink:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* This wraps the GEM_FLINK ioctl, which is used to export a gem buffer object
* into the device-global flink namespace. See gem_open() for opening such a
* buffer name on a different i915 drm file descriptor.
*
* Returns: The created flink buffer name.
*/
uint32_t gem_flink(int fd, uint32_t handle)
{
struct drm_gem_flink flink;
int ret;
memset(&flink, 0, sizeof(flink));
flink.handle = handle;
ret = ioctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
igt_assert(ret == 0);
errno = 0;
return flink.name;
}
/**
* gem_close:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This wraps the GEM_CLOSE ioctl, which to release a file-private gem buffer
* handle.
*/
void gem_close(int fd, uint32_t handle)
{
struct drm_gem_close close_bo;
igt_assert_neq(handle, 0);
memset(&close_bo, 0, sizeof(close_bo));
close_bo.handle = handle;
do_ioctl(fd, DRM_IOCTL_GEM_CLOSE, &close_bo);
}
static int __gem_write(int fd, uint32_t handle, uint64_t offset, const void *buf, uint64_t length)
{
struct drm_i915_gem_pwrite gem_pwrite;
int err;
memset(&gem_pwrite, 0, sizeof(gem_pwrite));
gem_pwrite.handle = handle;
gem_pwrite.offset = offset;
gem_pwrite.size = length;
gem_pwrite.data_ptr = (uintptr_t)buf;
err = 0;
if (drmIoctl(fd, DRM_IOCTL_I915_GEM_PWRITE, &gem_pwrite))
err = -errno;
return err;
}
/**
* gem_write:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset within the buffer of the subrange
* @buf: pointer to the data to write into the buffer
* @length: size of the subrange
*
* This wraps the PWRITE ioctl, which is to upload a linear data to a subrange
* of a gem buffer object.
*/
void gem_write(int fd, uint32_t handle, uint64_t offset, const void *buf, uint64_t length)
{
igt_assert_eq(__gem_write(fd, handle, offset, buf, length), 0);
}
static int __gem_read(int fd, uint32_t handle, uint64_t offset, void *buf, uint64_t length)
{
struct drm_i915_gem_pread gem_pread;
int err;
memset(&gem_pread, 0, sizeof(gem_pread));
gem_pread.handle = handle;
gem_pread.offset = offset;
gem_pread.size = length;
gem_pread.data_ptr = (uintptr_t)buf;
err = 0;
if (drmIoctl(fd, DRM_IOCTL_I915_GEM_PREAD, &gem_pread))
err = -errno;
return err;
}
/**
* gem_read:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset within the buffer of the subrange
* @buf: pointer to the data to read into
* @length: size of the subrange
*
* This wraps the PREAD ioctl, which is to download a linear data to a subrange
* of a gem buffer object.
*/
void gem_read(int fd, uint32_t handle, uint64_t offset, void *buf, uint64_t length)
{
igt_assert_eq(__gem_read(fd, handle, offset, buf, length), 0);
}
/**
* gem_set_domain:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @read_domains: gem domain bits for read access
* @write_domain: gem domain bit for write access
*
* This wraps the SET_DOMAIN ioctl, which is used to control the coherency of
* the gem buffer object between the cpu and gtt mappings. It is also use to
* synchronize with outstanding rendering in general, but for that use-case
* please have a look at gem_sync().
*/
void gem_set_domain(int fd, uint32_t handle,
uint32_t read_domains, uint32_t write_domain)
{
struct drm_i915_gem_set_domain set_domain;
memset(&set_domain, 0, sizeof(set_domain));
set_domain.handle = handle;
set_domain.read_domains = read_domains;
set_domain.write_domain = write_domain;
do_ioctl(fd, DRM_IOCTL_I915_GEM_SET_DOMAIN, &set_domain);
}
/**
* __gem_wait:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @timeout_ns: [in] time to wait, [out] remaining time (in nanoseconds)
*
* This functions waits for outstanding rendering to complete, upto
* the timeout_ns. If no timeout_ns is provided, the wait is indefinite and
* only returns upon an error or when the rendering is complete.
*/
int gem_wait(int fd, uint32_t handle, int64_t *timeout_ns)
{
struct drm_i915_gem_wait wait;
int ret;
memset(&wait, 0, sizeof(wait));
wait.bo_handle = handle;
wait.timeout_ns = timeout_ns ? *timeout_ns : -1;
wait.flags = 0;
ret = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_WAIT, &wait))
ret = -errno;
if (timeout_ns)
*timeout_ns = wait.timeout_ns;
return ret;
}
/**
* gem_sync:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This functions waits for outstanding rendering to complete.
*/
void gem_sync(int fd, uint32_t handle)
{
if (gem_wait(fd, handle, NULL))
gem_set_domain(fd, handle,
I915_GEM_DOMAIN_GTT,
I915_GEM_DOMAIN_GTT);
errno = 0;
}
bool gem_create__has_stolen_support(int fd)
{
static int has_stolen_support = -1;
struct drm_i915_getparam gp;
int val = -1;
if (has_stolen_support < 0) {
memset(&gp, 0, sizeof(gp));
gp.param = 38; /* CREATE_VERSION */
gp.value = &val;
/* Do we have the extended gem_create_ioctl? */
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp);
has_stolen_support = val >= 2;
}
return has_stolen_support;
}
struct local_i915_gem_create_v2 {
uint64_t size;
uint32_t handle;
uint32_t pad;
#define I915_CREATE_PLACEMENT_STOLEN (1<<0)
uint32_t flags;
};
#define LOCAL_IOCTL_I915_GEM_CREATE DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_CREATE, struct local_i915_gem_create_v2)
uint32_t __gem_create_stolen(int fd, uint64_t size)
{
struct local_i915_gem_create_v2 create;
int ret;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
create.flags = I915_CREATE_PLACEMENT_STOLEN;
ret = igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CREATE, &create);
if (ret < 0)
return 0;
errno = 0;
return create.handle;
}
/**
* gem_create_stolen:
* @fd: open i915 drm file descriptor
* @size: desired size of the buffer
*
* This wraps the new GEM_CREATE ioctl, which allocates a new gem buffer
* object of @size and placement in stolen memory region.
*
* Returns: The file-private handle of the created buffer object
*/
uint32_t gem_create_stolen(int fd, uint64_t size)
{
struct local_i915_gem_create_v2 create;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
create.flags = I915_CREATE_PLACEMENT_STOLEN;
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_CREATE, &create);
igt_assert(create.handle);
return create.handle;
}
uint32_t __gem_create(int fd, int size)
{
struct drm_i915_gem_create create;
int ret;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
ret = igt_ioctl(fd, DRM_IOCTL_I915_GEM_CREATE, &create);
if (ret < 0)
return 0;
errno = 0;
return create.handle;
}
/**
* gem_create:
* @fd: open i915 drm file descriptor
* @size: desired size of the buffer
*
* This wraps the GEM_CREATE ioctl, which allocates a new gem buffer object of
* @size.
*
* Returns: The file-private handle of the created buffer object
*/
uint32_t gem_create(int fd, uint64_t size)
{
struct drm_i915_gem_create create;
memset(&create, 0, sizeof(create));
create.handle = 0;
create.size = size;
do_ioctl(fd, DRM_IOCTL_I915_GEM_CREATE, &create);
igt_assert(create.handle);
return create.handle;
}
/**
* __gem_execbuf:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2 ioctl, which submits a batchbuffer for the gpu to
* run. This is allowed to fail, with -errno returned.
*/
int __gem_execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
int err = 0;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, execbuf))
err = -errno;
errno = 0;
return err;
}
/**
* gem_execbuf:
* @fd: open i915 drm file descriptor
* @execbuf: execbuffer data structure
*
* This wraps the EXECBUFFER2 ioctl, which submits a batchbuffer for the gpu to
* run.
*/
void gem_execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf)
{
igt_assert_eq(__gem_execbuf(fd, execbuf), 0);
}
/**
* __gem_mmap__gtt:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @size: size of the gem buffer
* @prot: memory protection bits as used by mmap()
*
* This functions wraps up procedure to establish a memory mapping through the
* GTT.
*
* Returns: A pointer to the created memory mapping, NULL on failure.
*/
void *__gem_mmap__gtt(int fd, uint32_t handle, uint64_t size, unsigned prot)
{
struct drm_i915_gem_mmap_gtt mmap_arg;
void *ptr;
memset(&mmap_arg, 0, sizeof(mmap_arg));
mmap_arg.handle = handle;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_MMAP_GTT, &mmap_arg))
return NULL;
ptr = mmap64(0, size, prot, MAP_SHARED, fd, mmap_arg.offset);
if (ptr == MAP_FAILED)
ptr = NULL;
else
errno = 0;
return ptr;
}
/**
* gem_mmap__gtt:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @size: size of the gem buffer
* @prot: memory protection bits as used by mmap()
*
* Like __gem_mmap__gtt() except we assert on failure.
*
* Returns: A pointer to the created memory mapping
*/
void *gem_mmap__gtt(int fd, uint32_t handle, uint64_t size, unsigned prot)
{
void *ptr = __gem_mmap__gtt(fd, handle, size, prot);
igt_assert(ptr);
return ptr;
}
struct local_i915_gem_mmap_v2 {
uint32_t handle;
uint32_t pad;
uint64_t offset;
uint64_t size;
uint64_t addr_ptr;
uint64_t flags;
#define I915_MMAP_WC 0x1
};
#define LOCAL_IOCTL_I915_GEM_MMAP_v2 DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_MMAP, struct local_i915_gem_mmap_v2)
bool gem_mmap__has_wc(int fd)
{
static int has_wc = -1;
if (has_wc == -1) {
struct drm_i915_getparam gp;
int val = -1;
has_wc = 0;
memset(&gp, 0, sizeof(gp));
gp.param = 30; /* MMAP_VERSION */
gp.value = &val;
/* Do we have the new mmap_ioctl? */
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp);
if (val >= 1) {
struct local_i915_gem_mmap_v2 arg;
/* Does this device support wc-mmaps ? */
memset(&arg, 0, sizeof(arg));
arg.handle = gem_create(fd, 4096);
arg.offset = 0;
arg.size = 4096;
arg.flags = I915_MMAP_WC;
has_wc = igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_MMAP_v2, &arg) == 0;
gem_close(fd, arg.handle);
}
errno = 0;
}
return has_wc > 0;
}
/**
* __gem_mmap__wc:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset in the gem buffer of the mmap arena
* @size: size of the mmap arena
* @prot: memory protection bits as used by mmap()
*
* This functions wraps up procedure to establish a memory mapping through
* direct cpu access, bypassing the gpu and cpu caches completely and also
* bypassing the GTT system agent (i.e. there is no automatic tiling of
* the mmapping through the fence registers).
*
* Returns: A pointer to the created memory mapping, NULL on failure.
*/
void *__gem_mmap__wc(int fd, uint32_t handle, uint64_t offset, uint64_t size, unsigned prot)
{
struct local_i915_gem_mmap_v2 arg;
if (!gem_mmap__has_wc(fd)) {
errno = ENOSYS;
return NULL;
}
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.offset = offset;
arg.size = size;
arg.flags = I915_MMAP_WC;
if (igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_MMAP_v2, &arg))
return NULL;
errno = 0;
return (void *)(uintptr_t)arg.addr_ptr;
}
/**
* gem_mmap__wc:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset in the gem buffer of the mmap arena
* @size: size of the mmap arena
* @prot: memory protection bits as used by mmap()
*
* Like __gem_mmap__wc() except we assert on failure.
*
* Returns: A pointer to the created memory mapping
*/
void *gem_mmap__wc(int fd, uint32_t handle, uint64_t offset, uint64_t size, unsigned prot)
{
void *ptr = __gem_mmap__wc(fd, handle, offset, size, prot);
igt_assert(ptr);
return ptr;
}
/**
* __gem_mmap__cpu:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset in the gem buffer of the mmap arena
* @size: size of the mmap arena
* @prot: memory protection bits as used by mmap()
*
* This functions wraps up procedure to establish a memory mapping through
* direct cpu access, bypassing the gpu completely.
*
* Returns: A pointer to the created memory mapping, NULL on failure.
*/
void *__gem_mmap__cpu(int fd, uint32_t handle, uint64_t offset, uint64_t size, unsigned prot)
{
struct drm_i915_gem_mmap mmap_arg;
memset(&mmap_arg, 0, sizeof(mmap_arg));
mmap_arg.handle = handle;
mmap_arg.offset = offset;
mmap_arg.size = size;
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_MMAP, &mmap_arg))
return NULL;
errno = 0;
return (void *)(uintptr_t)mmap_arg.addr_ptr;
}
/**
* gem_mmap__cpu:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @offset: offset in the gem buffer of the mmap arena
* @size: size of the mmap arena
* @prot: memory protection bits as used by mmap()
*
* Like __gem_mmap__cpu() except we assert on failure.
*
* Returns: A pointer to the created memory mapping
*/
void *gem_mmap__cpu(int fd, uint32_t handle, uint64_t offset, uint64_t size, unsigned prot)
{
void *ptr = __gem_mmap__cpu(fd, handle, offset, size, prot);
igt_assert(ptr);
return ptr;
}
/**
* gem_madvise:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
* @state: desired madvise state
*
* This is a wraps the MADVISE ioctl, which is used in libdrm to implement
* opportunistic buffer object caching. Objects in the cache are set to DONTNEED
* (internally in the kernel tracked as purgeable objects). When such a cached
* object is in need again it must be set back to WILLNEED before first use.
*
* Returns: When setting the madvise state to WILLNEED this returns whether the
* backing storage was still available or not.
*/
int gem_madvise(int fd, uint32_t handle, int state)
{
struct drm_i915_gem_madvise madv;
memset(&madv, 0, sizeof(madv));
madv.handle = handle;
madv.madv = state;
madv.retained = 1;
do_ioctl(fd, DRM_IOCTL_I915_GEM_MADVISE, &madv);
return madv.retained;
}
/**
* gem_context_create:
* @fd: open i915 drm file descriptor
*
* This is a wraps the CONTEXT_CREATE ioctl, which is used to allocate a new
* hardware context. Not that similarly to gem_set_caching() this wrapper calls
* igt_require() internally to correctly skip on kernels and platforms where hw
* context support is not available.
*
* Returns: The id of the allocated hw context.
*/
uint32_t gem_context_create(int fd)
{
struct drm_i915_gem_context_create create;
memset(&create, 0, sizeof(create));
if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_CREATE, &create)) {
int err = -errno;
igt_skip_on(err == -ENODEV || errno == -EINVAL);
igt_assert_eq(err, 0);
}
igt_assert(create.ctx_id != 0);
errno = 0;
return create.ctx_id;
}
int __gem_context_destroy(int fd, uint32_t ctx_id)
{
struct drm_i915_gem_context_destroy destroy;
int ret;
memset(&destroy, 0, sizeof(destroy));
destroy.ctx_id = ctx_id;
ret = igt_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_DESTROY, &destroy);
if (ret)
return -errno;
return 0;
}
/**
* gem_context_destroy:
* @fd: open i915 drm file descriptor
* @ctx_id: i915 hw context id
*
* This is a wraps the CONTEXT_DESTROY ioctl, which is used to free a hardware
* context.
*/
void gem_context_destroy(int fd, uint32_t ctx_id)
{
struct drm_i915_gem_context_destroy destroy;
memset(&destroy, 0, sizeof(destroy));
destroy.ctx_id = ctx_id;
do_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_DESTROY, &destroy);
}
int __gem_context_get_param(int fd, struct local_i915_gem_context_param *p)
{
#define LOCAL_I915_GEM_CONTEXT_GETPARAM 0x34
#define LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM DRM_IOWR (DRM_COMMAND_BASE + LOCAL_I915_GEM_CONTEXT_GETPARAM, struct local_i915_gem_context_param)
if (igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM, p))
return -errno;
errno = 0;
return 0;
}
/**
* gem_context_get_param:
* @fd: open i915 drm file descriptor
* @p: i915 hw context parameter
*
* This is a wraps the CONTEXT_GET_PARAM ioctl, which is used to free a hardware
* context. Not that similarly to gem_set_caching() this wrapper calls
* igt_require() internally to correctly skip on kernels and platforms where hw
* context parameter support is not available.
*/
void gem_context_get_param(int fd, struct local_i915_gem_context_param *p)
{
igt_assert(__gem_context_get_param(fd, p) == 0);
}
int __gem_context_set_param(int fd, struct local_i915_gem_context_param *p)
{
#define LOCAL_I915_GEM_CONTEXT_SETPARAM 0x35
#define LOCAL_IOCTL_I915_GEM_CONTEXT_SETPARAM DRM_IOWR (DRM_COMMAND_BASE + LOCAL_I915_GEM_CONTEXT_SETPARAM, struct local_i915_gem_context_param)
if (igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_SETPARAM, p))
return -errno;
errno = 0;
return 0;
}
/**
* gem_context_set_param:
* @fd: open i915 drm file descriptor
* @p: i915 hw context parameter
*
* This is a wraps the CONTEXT_SET_PARAM ioctl, which is used to free a hardware
* context. Not that similarly to gem_set_caching() this wrapper calls
* igt_require() internally to correctly skip on kernels and platforms where hw
* context parameter support is not available.
*/
void gem_context_set_param(int fd, struct local_i915_gem_context_param *p)
{
igt_assert(__gem_context_set_param(fd, p) == 0);
}
/**
* gem_context_require_param:
* @fd: open i915 drm file descriptor
* @param: i915 hw context parameter
*
* Feature test macro to query whether hw context parameter support for @param
* is available. Automatically skips through igt_require() if not.
*/
void gem_context_require_param(int fd, uint64_t param)
{
struct local_i915_gem_context_param p;
p.context = 0;
p.param = param;
p.value = 0;
p.size = 0;
igt_require(igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM, &p) == 0);
}
void gem_context_require_bannable(int fd)
{
static int has_ban_period = -1;
static int has_bannable = -1;
if (has_bannable < 0) {
struct local_i915_gem_context_param p;
p.context = 0;
p.param = LOCAL_CONTEXT_PARAM_BANNABLE;
p.value = 0;
p.size = 0;
has_bannable = igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM, &p) == 0;
}
if (has_ban_period < 0) {
struct local_i915_gem_context_param p;
p.context = 0;
p.param = LOCAL_CONTEXT_PARAM_BAN_PERIOD;
p.value = 0;
p.size = 0;
has_ban_period = igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM, &p) == 0;
}
igt_require(has_ban_period || has_bannable);
}
int __gem_userptr(int fd, void *ptr, int size, int read_only, uint32_t flags, uint32_t *handle)
{
struct local_i915_gem_userptr userptr;
memset(&userptr, 0, sizeof(userptr));
userptr.user_ptr = (uintptr_t)ptr;
userptr.user_size = size;
userptr.flags = flags;
if (read_only)
userptr.flags |= LOCAL_I915_USERPTR_READ_ONLY;
if (igt_ioctl(fd, LOCAL_IOCTL_I915_GEM_USERPTR, &userptr))
return -errno;
*handle = userptr.handle;
return 0;
}
/**
* gem_userptr:
* @fd: open i915 drm file descriptor
* @ptr: userptr pointer to be passed
* @size: desired size of the buffer
* @read_only: specify whether userptr is opened read only
* @flags: other userptr flags
* @handle: returned handle for the object
*
* Returns userptr handle for the GEM object.
*/
void gem_userptr(int fd, void *ptr, int size, int read_only, uint32_t flags, uint32_t *handle)
{
igt_assert_eq(__gem_userptr(fd, ptr, size, read_only, flags, handle), 0);
}
/**
* gem_sw_finish:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This is a wraps the SW_FINISH ioctl, which is used to flush out frontbuffer
* rendering done through the direct cpu memory mappings. Shipping userspace
* does _not_ call this after frontbuffer rendering through gtt memory mappings.
*/
void gem_sw_finish(int fd, uint32_t handle)
{
struct drm_i915_gem_sw_finish finish;
memset(&finish, 0, sizeof(finish));
finish.handle = handle;
do_ioctl(fd, DRM_IOCTL_I915_GEM_SW_FINISH, &finish);
}
/**
* gem_bo_busy:
* @fd: open i915 drm file descriptor
* @handle: gem buffer object handle
*
* This is a wraps the BUSY ioctl, which tells whether a buffer object is still
* actively used by the gpu in a execbuffer.
*
* Returns: The busy state of the buffer object.
*/
bool gem_bo_busy(int fd, uint32_t handle)
{
struct drm_i915_gem_busy busy;
memset(&busy, 0, sizeof(busy));
busy.handle = handle;
do_ioctl(fd, DRM_IOCTL_I915_GEM_BUSY, &busy);
return !!busy.busy;
}
/* feature test helpers */
/**
* gem_gtt_type:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check what type of gtt is being used by the kernel:
* 0 - global gtt
* 1 - aliasing ppgtt
* 2 - full ppgtt, limited to 32bit address space
* 3 - full ppgtt, 64bit address space
*
* Returns: Type of gtt being used.
*/
int gem_gtt_type(int fd)
{
struct drm_i915_getparam gp;
int val = 0;
memset(&gp, 0, sizeof(gp));
gp.param = 18; /* HAS_ALIASING_PPGTT */
gp.value = &val;
if (ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp)))
return 0;
errno = 0;
return val;
}
/**
* gem_uses_ppgtt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses ppgtt to
* execute batches. Note that this is also true when we're using full ppgtt.
*
* Returns: Whether batches are run through ppgtt.
*/
bool gem_uses_ppgtt(int fd)
{
return gem_gtt_type(fd) > 0;
}
/**
* gem_uses_full_ppgtt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to check whether the kernel internally uses full
* per-process gtt to execute batches. Note that this is also true when we're
* using full 64b ppgtt.
*
* Returns: Whether batches are run through full ppgtt.
*/
bool gem_uses_full_ppgtt(int fd)
{
return gem_gtt_type(fd) > 1;
}
/**
* gem_available_fences:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the number of available fences
* usable in a batchbuffer. Only relevant for pre-gen4.
*
* Returns: The number of available fences.
*/
int gem_available_fences(int fd)
{
static int num_fences = -1;
if (num_fences < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_NUM_FENCES_AVAIL;
gp.value = &num_fences;
num_fences = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return num_fences;
}
bool gem_has_llc(int fd)
{
static int has_llc = -1;
if (has_llc < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = I915_PARAM_HAS_LLC;
gp.value = &has_llc;
has_llc = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return has_llc;
}
static bool has_param(int fd, int param)
{
drm_i915_getparam_t gp;
int tmp = 0;
memset(&gp, 0, sizeof(gp));
gp.value = &tmp;
gp.param = param;
if (igt_ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp))
return false;
errno = 0;
return tmp > 0;
}
/**
* gem_has_bsd:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the BSD ring is available.
*
* Note that recent Bspec calls this the VCS ring for Video Command Submission.
*
* Returns: Whether the BSD ring is available or not.
*/
bool gem_has_bsd(int fd)
{
static int has_bsd = -1;
if (has_bsd < 0)
has_bsd = has_param(fd, I915_PARAM_HAS_BSD);
return has_bsd;
}
/**
* gem_has_blt:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the blitter ring is available.
*
* Note that recent Bspec calls this the BCS ring for Blitter Command Submission.
*
* Returns: Whether the blitter ring is available or not.
*/
bool gem_has_blt(int fd)
{
static int has_blt = -1;
if (has_blt < 0)
has_blt = has_param(fd, I915_PARAM_HAS_BLT);
return has_blt;
}
#define LOCAL_I915_PARAM_HAS_VEBOX 22
/**
* gem_has_vebox:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the vebox ring is available.
*
* Note that recent Bspec calls this the VECS ring for Video Enhancement Command
* Submission.
*
* Returns: Whether the vebox ring is available or not.
*/
bool gem_has_vebox(int fd)
{
static int has_vebox = -1;
if (has_vebox < 0)
has_vebox = has_param(fd, LOCAL_I915_PARAM_HAS_VEBOX);
return has_vebox;
}
#define LOCAL_I915_PARAM_HAS_BSD2 31
/**
* gem_has_bsd2:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the BSD2 ring is available.
*
* Note that recent Bspec calls this the VCS ring for Video Command Submission.
*
* Returns: Whether the BSD ring is avaible or not.
*/
bool gem_has_bsd2(int fd)
{
static int has_bsd2 = -1;
if (has_bsd2 < 0)
has_bsd2 = has_param(fd, LOCAL_I915_PARAM_HAS_BSD2);
return has_bsd2;
}
struct local_i915_gem_get_aperture {
__u64 aper_size;
__u64 aper_available_size;
__u64 version;
__u64 map_total_size;
__u64 stolen_total_size;
};
#define DRM_I915_GEM_GET_APERTURE 0x23
#define LOCAL_IOCTL_I915_GEM_GET_APERTURE DRM_IOR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_APERTURE, struct local_i915_gem_get_aperture)
/**
* gem_total_mappable_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total mappable size.
*
* Returns: Total mappable address space size.
*/
uint64_t gem_total_mappable_size(int fd)
{
struct local_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.map_total_size;
}
/**
* gem_total_stolen_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total stolen size.
*
* Returns: Total stolen memory.
*/
uint64_t gem_total_stolen_size(int fd)
{
struct local_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
do_ioctl(fd, LOCAL_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.stolen_total_size;
}
/**
* gem_available_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the available gpu aperture size
* usable in a batchbuffer.
*
* Returns: The available gtt address space size.
*/
uint64_t gem_available_aperture_size(int fd)
{
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.aper_available_size;
}
/**
* gem_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the total gpu aperture size.
*
* Returns: The total gtt address space size.
*/
uint64_t gem_aperture_size(int fd)
{
static uint64_t aperture_size = 0;
if (aperture_size == 0) {
struct local_i915_gem_context_param p;
memset(&p, 0, sizeof(p));
p.param = 0x3;
if (ioctl(fd, LOCAL_IOCTL_I915_GEM_CONTEXT_GETPARAM, &p) == 0) {
aperture_size = p.value;
} else {
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
aperture_size = aperture.aper_size;
}
}
return aperture_size;
}
/**
* gem_mappable_aperture_size:
*
* Feature test macro to query the kernel for the mappable gpu aperture size.
* This is the area available for GTT memory mappings.
*
* Returns: The mappable gtt address space size.
*/
uint64_t gem_mappable_aperture_size(void)
{
struct pci_device *pci_dev = intel_get_pci_device();
int bar;
if (intel_gen(pci_dev->device_id) < 3)
bar = 0;
else
bar = 2;
return pci_dev->regions[bar].size;
}
/**
* gem_global_aperture_size:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query the kernel for the global gpu aperture size.
* This is the area available for the kernel to perform address translations.
*
* Returns: The mappable gtt address space size.
*/
uint64_t gem_global_aperture_size(int fd)
{
struct drm_i915_gem_get_aperture aperture;
memset(&aperture, 0, sizeof(aperture));
aperture.aper_size = 256*1024*1024;
do_ioctl(fd, DRM_IOCTL_I915_GEM_GET_APERTURE, &aperture);
return aperture.aper_size;
}
#define LOCAL_I915_PARAM_HAS_EXEC_SOFTPIN 37
/**
* gem_has_softpin:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the softpinning functionality is
* supported.
*
* Returns: Whether softpin support is available
*/
bool gem_has_softpin(int fd)
{
static int has_softpin = -1;
if (has_softpin < 0) {
struct drm_i915_getparam gp;
memset(&gp, 0, sizeof(gp));
gp.param = LOCAL_I915_PARAM_HAS_EXEC_SOFTPIN;
gp.value = &has_softpin;
has_softpin = 0;
ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp, sizeof(gp));
errno = 0;
}
return has_softpin;
}
/**
* gem_require_caching:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether buffer object caching control is
* available. Automatically skips through igt_require() if not.
*/
void gem_require_caching(int fd)
{
uint32_t handle;
handle = gem_create(fd, 4096);
gem_set_caching(fd, handle, 0);
gem_close(fd, handle);
errno = 0;
}
bool gem_has_ring(int fd, unsigned ring)
{
struct drm_i915_gem_execbuffer2 execbuf;
struct drm_i915_gem_exec_object2 exec;
/* silly ABI, the kernel thinks everyone who has BSD also has BSD2 */
if ((ring & ~(3<<13)) == I915_EXEC_BSD) {
if (ring & (3 << 13) && !gem_has_bsd2(fd))
return false;
}
memset(&exec, 0, sizeof(exec));
memset(&execbuf, 0, sizeof(execbuf));
execbuf.buffers_ptr = (uintptr_t)&exec;
execbuf.buffer_count = 1;
execbuf.flags = ring;
return __gem_execbuf(fd, &execbuf) == -ENOENT;
}
/**
* gem_require_ring:
* @fd: open i915 drm file descriptor
* @ring: ring flag bit as used in gem_execbuf()
*
* Feature test macro to query whether a specific ring is available.
* This automagically skips if the ring isn't available by
* calling igt_require().
*/
void gem_require_ring(int fd, unsigned ring)
{
igt_require(gem_has_ring(fd, ring));
}
/**
* gem_has_mocs_registers:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the device has MOCS registers.
* These exist gen 9+.
*/
bool gem_has_mocs_registers(int fd)
{
return intel_gen(intel_get_drm_devid(fd)) >= 9;
}
/**
* gem_require_mocs_registers:
* @fd: open i915 drm file descriptor
*
* Feature test macro to query whether the device has MOCS registers.
* These exist gen 9+.
*/
void gem_require_mocs_registers(int fd)
{
igt_require(gem_has_mocs_registers(fd));
}
/* prime */
/**
* prime_handle_to_fd:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* This wraps the PRIME_HANDLE_TO_FD ioctl, which is used to export a gem buffer
* object into a global (i.e. potentially cross-device) dma-buf file-descriptor
* handle.
*
* Returns: The created dma-buf fd handle.
*/
int prime_handle_to_fd(int fd, uint32_t handle)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.handle = handle;
args.flags = DRM_CLOEXEC;
args.fd = -1;
do_ioctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args);
return args.fd;
}
/**
* prime_handle_to_fd_for_mmap:
* @fd: open i915 drm file descriptor
* @handle: file-private gem buffer object handle
*
* Same as prime_handle_to_fd above but with DRM_RDWR capabilities, which can
* be useful for writing into the mmap'ed dma-buf file-descriptor.
*
* Returns: The created dma-buf fd handle or -1 if the ioctl fails.
*/
int prime_handle_to_fd_for_mmap(int fd, uint32_t handle)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.handle = handle;
args.flags = DRM_CLOEXEC | DRM_RDWR;
args.fd = -1;
if (igt_ioctl(fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &args) != 0)
return -1;
return args.fd;
}
/**
* prime_fd_to_handle:
* @fd: open i915 drm file descriptor
* @dma_buf_fd: dma-buf fd handle
*
* This wraps the PRIME_FD_TO_HANDLE ioctl, which is used to import a dma-buf
* file-descriptor into a gem buffer object.
*
* Returns: The created gem buffer object handle.
*/
uint32_t prime_fd_to_handle(int fd, int dma_buf_fd)
{
struct drm_prime_handle args;
memset(&args, 0, sizeof(args));
args.fd = dma_buf_fd;
args.flags = 0;
args.handle = 0;
do_ioctl(fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &args);
return args.handle;
}
/**
* prime_get_size:
* @dma_buf_fd: dma-buf fd handle
*
* This wraps the lseek() protocol used to query the invariant size of a
* dma-buf. Not all kernels support this, which is check with igt_require() and
* so will result in automagic test skipping.
*
* Returns: The lifetime-invariant size of the dma-buf object.
*/
off_t prime_get_size(int dma_buf_fd)
{
off_t ret;
ret = lseek(dma_buf_fd, 0, SEEK_END);
igt_assert(ret >= 0 || errno == ESPIPE);
igt_require(ret >= 0);
errno = 0;
return ret;
}
/**
* prime_sync_start
* @dma_buf_fd: dma-buf fd handle
* @write: read/write or read-only access
*
* Must be called before starting CPU mmap access to a dma-buf.
*/
void prime_sync_start(int dma_buf_fd, bool write)
{
struct local_dma_buf_sync sync_start;
memset(&sync_start, 0, sizeof(sync_start));
sync_start.flags = LOCAL_DMA_BUF_SYNC_START;
sync_start.flags |= LOCAL_DMA_BUF_SYNC_READ;
if (write)
sync_start.flags |= LOCAL_DMA_BUF_SYNC_WRITE;
do_ioctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync_start);
}
/**
* prime_sync_end
* @dma_buf_fd: dma-buf fd handle
* @write: read/write or read-only access
*
* Must be called after finishing CPU mmap access to a dma-buf.
*/
void prime_sync_end(int dma_buf_fd, bool write)
{
struct local_dma_buf_sync sync_end;
memset(&sync_end, 0, sizeof(sync_end));
sync_end.flags = LOCAL_DMA_BUF_SYNC_END;
sync_end.flags |= LOCAL_DMA_BUF_SYNC_READ;
if (write)
sync_end.flags |= LOCAL_DMA_BUF_SYNC_WRITE;
do_ioctl(dma_buf_fd, LOCAL_DMA_BUF_IOCTL_SYNC, &sync_end);
}
/**
* igt_require_fb_modifiers:
* @fd: Open DRM file descriptor.
*
* Requires presence of DRM_CAP_ADDFB2_MODIFIERS.
*/
void igt_require_fb_modifiers(int fd)
{
static bool has_modifiers, cap_modifiers_tested;
if (!cap_modifiers_tested) {
uint64_t cap_modifiers;
int ret;
ret = drmGetCap(fd, LOCAL_DRM_CAP_ADDFB2_MODIFIERS, &cap_modifiers);
igt_assert(ret == 0 || errno == EINVAL);
has_modifiers = ret == 0 && cap_modifiers == 1;
cap_modifiers_tested = true;
}
igt_require(has_modifiers);
}
int __kms_addfb(int fd, uint32_t handle, uint32_t width, uint32_t height,
uint32_t stride, uint32_t pixel_format, uint64_t modifier,
uint32_t flags, uint32_t *buf_id)
{
struct local_drm_mode_fb_cmd2 f;
int ret;
igt_require_fb_modifiers(fd);
memset(&f, 0, sizeof(f));
f.width = width;
f.height = height;
f.pixel_format = pixel_format;
f.flags = flags;
f.handles[0] = handle;
f.pitches[0] = stride;
f.modifier[0] = modifier;
ret = igt_ioctl(fd, LOCAL_DRM_IOCTL_MODE_ADDFB2, &f);
*buf_id = f.fb_id;
return ret < 0 ? -errno : ret;
}