blob: a14a8e16393781722171313e2a49b0a5d03117f5 [file] [log] [blame]
/*
* XGL
*
* Copyright (C) 2014 LunarG, Inc.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included
* in all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*
* Authors:
* Chia-I Wu <olv@lunarg.com>
*/
#include <stdio.h>
#include <sys/types.h>
#include <sys/stat.h>
#include <fcntl.h>
#include <unistd.h>
#include "genhw/genhw.h"
#include "dispatch.h"
#include "queue.h"
#include "gpu.h"
static const char *gpu_get_name(const struct intel_gpu *gpu)
{
const char *name = NULL;
if (gen_is_hsw(gpu->devid)) {
if (gen_is_desktop(gpu->devid))
name = "Intel(R) Haswell Desktop";
else if (gen_is_mobile(gpu->devid))
name = "Intel(R) Haswell Mobile";
else if (gen_is_server(gpu->devid))
name = "Intel(R) Haswell Server";
}
else if (gen_is_ivb(gpu->devid)) {
if (gen_is_desktop(gpu->devid))
name = "Intel(R) Ivybridge Desktop";
else if (gen_is_mobile(gpu->devid))
name = "Intel(R) Ivybridge Mobile";
else if (gen_is_server(gpu->devid))
name = "Intel(R) Ivybridge Server";
}
else if (gen_is_snb(gpu->devid)) {
if (gen_is_desktop(gpu->devid))
name = "Intel(R) Sandybridge Desktop";
else if (gen_is_mobile(gpu->devid))
name = "Intel(R) Sandybridge Mobile";
else if (gen_is_server(gpu->devid))
name = "Intel(R) Sandybridge Server";
}
if (!name)
name = "Unknown Intel Chipset";
return name;
}
static int gpu_open_internal(struct intel_gpu *gpu)
{
if (gpu->fd_internal < 0) {
gpu->fd_internal = open(gpu->path, O_RDWR);
if (gpu->fd_internal < 0) {
icd_log(XGL_DBG_MSG_ERROR, XGL_VALIDATION_LEVEL_0, NULL, 0,
0, "failed to open %s", gpu->path);
}
}
return gpu->fd_internal;
}
static void gpu_close_internal(struct intel_gpu *gpu)
{
if (gpu->fd_internal >= 0) {
close(gpu->fd_internal);
gpu->fd_internal = -1;
}
}
static struct intel_gpu *gpu_create(int gen, int devid, const char *path)
{
struct intel_gpu *gpu;
size_t path_len;
gpu = icd_alloc(sizeof(*gpu), 0, XGL_SYSTEM_ALLOC_API_OBJECT);
if (!gpu)
return NULL;
memset(gpu, 0, sizeof(*gpu));
/* debug layer is always enabled for intel_gpu */
gpu->dispatch = intel_dispatch_get(true);
gpu->devid = devid;
path_len = strlen(path);
gpu->path = icd_alloc(path_len + 1, 0, XGL_SYSTEM_ALLOC_INTERNAL);
if (!gpu->path) {
icd_free(gpu);
return NULL;
}
memcpy(gpu->path, path, path_len + 1);
gpu->gen_opaque = gen;
switch (intel_gpu_gen(gpu)) {
case INTEL_GEN(7.5):
gpu->gt = gen_get_hsw_gt(devid);
if (gpu->gt == 3) {
gpu->max_vs_threads = 280;
gpu->max_gs_threads = 256;
gpu->max_fs_threads = 408;
} else if (gpu->gt == 2){
gpu->max_vs_threads = 280;
gpu->max_gs_threads = 256;
gpu->max_fs_threads = 204;
} else {
gpu->max_vs_threads = 70;
gpu->max_gs_threads = 70;
gpu->max_fs_threads = 102;
}
break;
case INTEL_GEN(7):
gpu->gt = gen_get_ivb_gt(devid);
if (gpu->gt == 2) {
gpu->max_vs_threads = 128;
gpu->max_gs_threads = 128;
gpu->max_fs_threads = 172;
} else {
gpu->max_vs_threads = 36;
gpu->max_gs_threads = 36;
gpu->max_fs_threads = 48;
}
break;
case INTEL_GEN(6):
gpu->gt = gen_get_snb_gt(devid);
if (gpu->gt == 2) {
gpu->max_vs_threads = 60;
gpu->max_gs_threads = 60;
gpu->max_fs_threads = 80;
} else {
gpu->max_vs_threads = 24;
gpu->max_gs_threads = 21;
gpu->max_fs_threads = 40;
}
break;
}
/* 8192 dwords */
gpu->max_batch_buffer_size = sizeof(uint32_t) * 8192;
/* the winsys is prepared for one reloc every two dwords, then minus 2 */
gpu->batch_buffer_reloc_count =
gpu->max_batch_buffer_size / sizeof(uint32_t) / 2 - 2;
gpu->fd_internal = -1;
gpu->fd = -1;
return gpu;
}
static void gpu_destroy(struct intel_gpu *gpu)
{
gpu_close_internal(gpu);
icd_free(gpu->path);
icd_free(gpu);
}
static struct intel_gpu *intel_gpus;
/**
* Return true if \p gpu is a valid intel_gpu.
*/
bool intel_gpu_is_valid(const struct intel_gpu *gpu)
{
const struct intel_gpu *iter = intel_gpus;
while (iter) {
if (iter == gpu)
return true;
iter = iter->next;
}
return false;
}
static int devid_to_gen(int devid)
{
int gen;
if (gen_is_hsw(devid))
gen = INTEL_GEN(7.5);
else if (gen_is_ivb(devid))
gen = INTEL_GEN(7);
else if (gen_is_snb(devid))
gen = INTEL_GEN(6);
else
gen = -1;
#ifdef INTEL_GEN_SPECIALIZED
if (gen != INTEL_GEN(INTEL_GEN_SPECIALIZED))
gen = -1;
#endif
return gen;
}
XGL_RESULT intel_gpu_add(int devid, const char *path,
struct intel_gpu **gpu_ret)
{
const int gen = devid_to_gen(devid);
struct intel_gpu *gpu;
if (gen < 0) {
icd_log(XGL_DBG_MSG_WARNING, XGL_VALIDATION_LEVEL_0, XGL_NULL_HANDLE,
0, 0, "unsupported device id 0x%04x", devid);
return XGL_ERROR_INITIALIZATION_FAILED;
}
gpu = gpu_create(gen, devid, path);
if (!gpu)
return XGL_ERROR_OUT_OF_MEMORY;
gpu->next = intel_gpus;
intel_gpus = gpu;
*gpu_ret = gpu;
return XGL_SUCCESS;
}
void intel_gpu_remove_all(void)
{
struct intel_gpu *gpu = intel_gpus;
while (gpu) {
struct intel_gpu *next = gpu->next;
gpu_destroy(gpu);
gpu = next;
}
intel_gpus = NULL;
}
struct intel_gpu *intel_gpu_get_list(void)
{
return intel_gpus;
}
void intel_gpu_get_props(const struct intel_gpu *gpu,
XGL_PHYSICAL_GPU_PROPERTIES *props)
{
const char *name;
size_t name_len;
props->structSize = sizeof(*props);
props->apiVersion = INTEL_API_VERSION;
props->driverVersion = INTEL_DRIVER_VERSION;
props->vendorId = 0x8086;
props->deviceId = gpu->devid;
props->gpuType = XGL_GPU_TYPE_INTEGRATED;
/* copy GPU name */
name = gpu_get_name(gpu);
name_len = strlen(name);
if (name_len > sizeof(props->gpuName) - 1)
name_len = sizeof(props->gpuName) - 1;
memcpy(props->gpuName, name, name_len);
props->gpuName[name_len] = '\0';
props->maxMemRefsPerSubmission = gpu->batch_buffer_reloc_count;
props->virtualMemPageSize = 4096;
/* no size limit, but no bounded buffer could exceed 2GB */
props->maxInlineMemoryUpdateSize = 2u << 30;
props->maxBoundDescriptorSets = 1;
props->maxThreadGroupSize = 512;
/* incremented every 80ns */
props->timestampFrequency = 1000 * 1000 * 1000 / 80;
props->multiColorAttachmentClears = false;
}
void intel_gpu_get_perf(const struct intel_gpu *gpu,
XGL_PHYSICAL_GPU_PERFORMANCE *perf)
{
/* TODO */
perf->maxGpuClock = 1.0f;
perf->aluPerClock = 1.0f;
perf->texPerClock = 1.0f;
perf->primsPerClock = 1.0f;
perf->pixelsPerClock = 1.0f;
}
void intel_gpu_get_queue_props(const struct intel_gpu *gpu,
enum intel_gpu_engine_type engine,
XGL_PHYSICAL_GPU_QUEUE_PROPERTIES *props)
{
props->structSize = sizeof(*props);
switch (engine) {
case INTEL_GPU_ENGINE_3D:
props->queueFlags = XGL_QUEUE_GRAPHICS_BIT | XGL_QUEUE_COMPUTE_BIT;
props->queueCount = 1;
props->maxAtomicCounters = INTEL_QUEUE_ATOMIC_COUNTER_COUNT;
props->supportsTimestamps = true;
break;
default:
assert(!"unknown engine type");
return;
}
}
void intel_gpu_get_memory_props(const struct intel_gpu *gpu,
XGL_PHYSICAL_GPU_MEMORY_PROPERTIES *props)
{
props->structSize = sizeof(*props);
props->supportsMigration = false;
/* no kernel support yet */
props->supportsVirtualMemoryRemapping = false;
/* no winsys support for DRM_I915_GEM_USERPTR yet */
props->supportsPinning = false;
}
XGL_RESULT intel_gpu_open(struct intel_gpu *gpu)
{
gpu->fd = gpu_open_internal(gpu);
return (gpu->fd >= 0) ? XGL_SUCCESS : XGL_ERROR_UNKNOWN;
}
void intel_gpu_close(struct intel_gpu *gpu)
{
gpu->fd = -1;
gpu_close_internal(gpu);
}
bool intel_gpu_has_extension(const struct intel_gpu *gpu, const char *ext)
{
return false;
}
XGL_RESULT XGLAPI intelGetGpuInfo(
XGL_PHYSICAL_GPU gpu_,
XGL_PHYSICAL_GPU_INFO_TYPE infoType,
XGL_SIZE* pDataSize,
XGL_VOID* pData)
{
const struct intel_gpu *gpu = intel_gpu(gpu_);
XGL_RESULT ret = XGL_SUCCESS;
switch (infoType) {
case XGL_INFO_TYPE_PHYSICAL_GPU_PROPERTIES:
if (pData == NULL) {
return XGL_ERROR_INVALID_POINTER;
}
*pDataSize = sizeof(XGL_PHYSICAL_GPU_PROPERTIES);
intel_gpu_get_props(gpu, pData);
break;
case XGL_INFO_TYPE_PHYSICAL_GPU_PERFORMANCE:
if (pData == NULL) {
return XGL_ERROR_INVALID_POINTER;
}
*pDataSize = sizeof(XGL_PHYSICAL_GPU_PERFORMANCE);
intel_gpu_get_perf(gpu, pData);
break;
case XGL_INFO_TYPE_PHYSICAL_GPU_QUEUE_PROPERTIES:
/*
* XGL Programmers guide, page 33:
* to determine the data size an application calls
* xglGetGpuInfo() with a NULL data pointer. The
* expected data size for all queue property structures
* is returned in pDataSize
*/
*pDataSize = sizeof(XGL_PHYSICAL_GPU_QUEUE_PROPERTIES) *
INTEL_GPU_ENGINE_COUNT;
if (pData != NULL) {
XGL_PHYSICAL_GPU_QUEUE_PROPERTIES *dst = pData;
int engine;
for (engine = 0; engine < INTEL_GPU_ENGINE_COUNT; engine++) {
intel_gpu_get_queue_props(gpu, engine, dst);
dst++;
}
}
break;
case XGL_INFO_TYPE_PHYSICAL_GPU_MEMORY_PROPERTIES:
if (pData == NULL) {
return XGL_ERROR_INVALID_POINTER;
}
*pDataSize = sizeof(XGL_PHYSICAL_GPU_MEMORY_PROPERTIES);
intel_gpu_get_memory_props(gpu, pData);
break;
default:
ret = XGL_ERROR_INVALID_VALUE;
}
return ret;
}
XGL_RESULT XGLAPI intelGetExtensionSupport(
XGL_PHYSICAL_GPU gpu_,
const XGL_CHAR* pExtName)
{
struct intel_gpu *gpu = intel_gpu(gpu_);
return (intel_gpu_has_extension(gpu, (const char *) pExtName)) ?
XGL_SUCCESS : XGL_ERROR_INVALID_EXTENSION;
}
XGL_RESULT XGLAPI intelGetMultiGpuCompatibility(
XGL_PHYSICAL_GPU gpu0_,
XGL_PHYSICAL_GPU gpu1_,
XGL_GPU_COMPATIBILITY_INFO* pInfo)
{
const struct intel_gpu *gpu0 = intel_gpu(gpu0_);
const struct intel_gpu *gpu1 = intel_gpu(gpu1_);
XGL_FLAGS compat = XGL_GPU_COMPAT_IQ_MATCH_BIT |
XGL_GPU_COMPAT_PEER_TRANSFER_BIT |
XGL_GPU_COMPAT_SHARED_MEMORY_BIT |
XGL_GPU_COMPAT_SHARED_GPU0_DISPLAY_BIT |
XGL_GPU_COMPAT_SHARED_GPU1_DISPLAY_BIT;
if (intel_gpu_gen(gpu0) == intel_gpu_gen(gpu1))
compat |= XGL_GPU_COMPAT_ASIC_FEATURES_BIT;
pInfo->compatibilityFlags = compat;
return XGL_SUCCESS;
}