blob: 943f58dac89740f0808505993c269ac3d13bdaee [file] [log] [blame]
/*
* Copyright © 2016 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
* IN THE SOFTWARE.
*/
#include "igt.h"
#include "igt_rand.h"
IGT_TEST_DESCRIPTION("Fill the GTT with batches.");
#define BATCH_SIZE (4096<<10)
static void gem_require_store_dword(int fd, unsigned ring)
{
int gen = intel_gen(intel_get_drm_devid(fd));
ring &= ~(3 << 13);
igt_skip_on_f(gen == 6 && ring == I915_EXEC_BSD,
"MI_STORE_DATA broken on gen6 bsd\n");
}
static bool ignore_engine(int fd, unsigned engine)
{
if (engine == 0)
return true;
if (gem_has_bsd2(fd) && engine == I915_EXEC_BSD)
return true;
return false;
}
static void xchg_u32(void *array, unsigned i, unsigned j)
{
uint32_t *u32 = array;
uint32_t tmp = u32[i];
u32[i] = u32[j];
u32[j] = tmp;
}
static void submit(int fd, int gen,
struct drm_i915_gem_execbuffer2 *eb,
struct drm_i915_gem_relocation_entry *reloc,
uint32_t *handles, unsigned count)
{
struct drm_i915_gem_exec_object2 obj;
uint32_t batch[16];
unsigned n;
memset(&obj, 0, sizeof(obj));
obj.relocs_ptr = (uintptr_t)reloc;
obj.relocation_count = 2;
memset(reloc, 0, 2*sizeof(*reloc));
reloc[0].offset = eb->batch_start_offset;
reloc[0].offset += sizeof(uint32_t);
reloc[0].delta = BATCH_SIZE - eb->batch_start_offset - 8;
reloc[0].read_domains = I915_GEM_DOMAIN_INSTRUCTION;
reloc[1].offset = eb->batch_start_offset;
reloc[1].offset += 3*sizeof(uint32_t);
reloc[1].read_domains = I915_GEM_DOMAIN_INSTRUCTION;
n = 0;
batch[n] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0);
if (gen >= 8) {
batch[n] |= 1 << 21;
batch[n]++;
batch[++n] = reloc[0].delta;/* lower_32_bits(address) */
batch[++n] = 0; /* upper_32_bits(address) */
} else if (gen >= 4) {
batch[++n] = 0;
batch[++n] = reloc[0].delta;/* lower_32_bits(address) */
reloc[0].offset += sizeof(uint32_t);
} else {
batch[n]--;
batch[++n] = reloc[0].delta;/* lower_32_bits(address) */
reloc[1].offset -= sizeof(uint32_t);
}
batch[++n] = 0; /* lower_32_bits(value) */
batch[++n] = 0; /* upper_32_bits(value) / nop */
batch[++n] = MI_BATCH_BUFFER_END;
eb->buffers_ptr = (uintptr_t)&obj;
for (unsigned i = 0; i < count; i++) {
obj.handle = handles[i];
reloc[0].target_handle = obj.handle;
reloc[1].target_handle = obj.handle;
obj.offset = 0;
reloc[0].presumed_offset = obj.offset;
reloc[1].presumed_offset = obj.offset;
gem_write(fd, obj.handle, eb->batch_start_offset,
batch, sizeof(batch));
gem_execbuf(fd, eb);
}
/* As we have been lying about the write_domain, we need to do a sync */
gem_sync(fd, obj.handle);
}
static void fillgtt(int fd, unsigned ring, int timeout)
{
const int gen = intel_gen(intel_get_drm_devid(fd));
struct drm_i915_gem_execbuffer2 execbuf;
struct drm_i915_gem_relocation_entry reloc[2];
volatile uint64_t *shared;
unsigned *handles;
unsigned engines[16];
unsigned nengine;
unsigned engine;
uint64_t size;
unsigned count;
shared = mmap(NULL, 4096, PROT_WRITE, MAP_SHARED | MAP_ANON, -1, 0);
igt_assert(shared != MAP_FAILED);
nengine = 0;
if (ring == 0) {
for_each_engine(fd, engine) {
if (ignore_engine(fd, engine))
continue;
if (gen == 6 && e__->exec_id == I915_EXEC_BSD)
continue;
engines[nengine++] = engine;
}
} else {
gem_require_ring(fd, ring);
gem_require_store_dword(fd, ring);
engines[nengine++] = ring;
}
igt_require(nengine);
size = gem_aperture_size(fd);
if (size > 1ull<<32) /* Limit to 4GiB as we do not use allow-48b */
size = 1ull << 32;
igt_require(size < (1ull<<32) * BATCH_SIZE);
count = size / BATCH_SIZE + 1;
igt_debug("Using %'d batches to fill %'llu aperture on %d engines\n",
count, (long long)size, nengine);
intel_require_memory(count, BATCH_SIZE, CHECK_RAM);
intel_detect_and_clear_missed_interrupts(fd);
memset(&execbuf, 0, sizeof(execbuf));
execbuf.buffer_count = 1;
if (gen < 6)
execbuf.flags |= I915_EXEC_SECURE;
handles = calloc(count, sizeof(handles));
igt_assert(handles);
for (unsigned i = 0; i < count; i++)
handles[i] = gem_create(fd, BATCH_SIZE);
/* Flush all memory before we start the timer */
submit(fd, gen, &execbuf, reloc, handles, count);
igt_fork(child, nengine) {
uint64_t cycles = 0;
hars_petruska_f54_1_random_perturb(child);
igt_permute_array(handles, count, xchg_u32);
execbuf.batch_start_offset = child*64;
execbuf.flags |= engines[child];
igt_until_timeout(timeout) {
submit(fd, gen, &execbuf, reloc, handles, count);
for (unsigned i = 0; i < count; i++) {
uint32_t handle = handles[i];
uint64_t buf[2];
gem_read(fd, handle, reloc[1].offset, &buf[0], sizeof(buf[0]));
gem_read(fd, handle, reloc[0].delta, &buf[1], sizeof(buf[1]));
igt_assert_eq_u64(buf[0], buf[1]);
}
cycles++;
}
shared[child] = cycles;
igt_info("engine[%d]: %llu cycles\n", child, (long long)cycles);
}
igt_waitchildren();
for (unsigned i = 0; i < count; i++)
gem_close(fd, handles[i]);
shared[nengine] = 0;
for (unsigned i = 0; i < nengine; i++)
shared[nengine] += shared[i];
igt_info("Total: %llu cycles\n", (long long)shared[nengine]);
igt_assert_eq(intel_detect_and_clear_missed_interrupts(fd), 0);
}
igt_main
{
const struct intel_execution_engine *e;
int device = -1;
igt_skip_on_simulation();
igt_fixture {
device = drm_open_driver(DRIVER_INTEL);
igt_fork_hang_detector(device);
}
igt_subtest("basic")
fillgtt(device, 0, 1); /* just enough to run a single pass */
for (e = intel_execution_engines; e->name; e++)
igt_subtest_f("%s", e->name)
fillgtt(device, e->exec_id | e->flags, 20);
igt_subtest("all")
fillgtt(device, 0, 150);
igt_fixture {
igt_stop_hang_detector();
close(device);
}
}