Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright © 2008 Intel Corporation |
| 3 | * |
| 4 | * Permission is hereby granted, free of charge, to any person obtaining a |
| 5 | * copy of this software and associated documentation files (the "Software"), |
| 6 | * to deal in the Software without restriction, including without limitation |
| 7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| 8 | * and/or sell copies of the Software, and to permit persons to whom the |
| 9 | * Software is furnished to do so, subject to the following conditions: |
| 10 | * |
| 11 | * The above copyright notice and this permission notice (including the next |
| 12 | * paragraph) shall be included in all copies or substantial portions of the |
| 13 | * Software. |
| 14 | * |
| 15 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| 16 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| 17 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| 18 | * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| 19 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| 20 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
| 21 | * IN THE SOFTWARE. |
| 22 | * |
| 23 | * Authors: |
| 24 | * Eric Anholt <eric@anholt.net> |
| 25 | * |
| 26 | */ |
| 27 | |
David Howells | 760285e | 2012-10-02 18:01:07 +0100 | [diff] [blame] | 28 | #include <linux/string.h> |
| 29 | #include <linux/bitops.h> |
| 30 | #include <drm/drmP.h> |
| 31 | #include <drm/i915_drm.h> |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 32 | #include "i915_drv.h" |
| 33 | |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 34 | /** |
| 35 | * DOC: buffer object tiling |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 36 | * |
Chris Wilson | 111dbca | 2017-01-10 12:10:44 +0000 | [diff] [blame^] | 37 | * i915_gem_set_tiling_ioctl() and i915_gem_get_tiling_ioctl() is the userspace |
| 38 | * interface to declare fence register requirements. |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 39 | * |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 40 | * In principle GEM doesn't care at all about the internal data layout of an |
| 41 | * object, and hence it also doesn't care about tiling or swizzling. There's two |
| 42 | * exceptions: |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 43 | * |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 44 | * - For X and Y tiling the hardware provides detilers for CPU access, so called |
| 45 | * fences. Since there's only a limited amount of them the kernel must manage |
| 46 | * these, and therefore userspace must tell the kernel the object tiling if it |
| 47 | * wants to use fences for detiling. |
| 48 | * - On gen3 and gen4 platforms have a swizzling pattern for tiled objects which |
| 49 | * depends upon the physical page frame number. When swapping such objects the |
| 50 | * page frame number might change and the kernel must be able to fix this up |
| 51 | * and hence now the tiling. Note that on a subset of platforms with |
| 52 | * asymmetric memory channel population the swizzling pattern changes in an |
| 53 | * unknown way, and for those the kernel simply forbids swapping completely. |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 54 | * |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 55 | * Since neither of this applies for new tiling layouts on modern platforms like |
| 56 | * W, Ys and Yf tiling GEM only allows object tiling to be set to X or Y tiled. |
| 57 | * Anything else can be handled in userspace entirely without the kernel's |
| 58 | * invovlement. |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 59 | */ |
| 60 | |
Chris Wilson | 91d4e0aa | 2017-01-09 16:16:13 +0000 | [diff] [blame] | 61 | /** |
| 62 | * i915_gem_fence_size - required global GTT size for a fence |
| 63 | * @i915: i915 device |
| 64 | * @size: object size |
| 65 | * @tiling: tiling mode |
| 66 | * @stride: tiling stride |
| 67 | * |
| 68 | * Return the required global GTT size for a fence (view of a tiled object), |
| 69 | * taking into account potential fence register mapping. |
| 70 | */ |
| 71 | u32 i915_gem_fence_size(struct drm_i915_private *i915, |
| 72 | u32 size, unsigned int tiling, unsigned int stride) |
| 73 | { |
| 74 | u32 ggtt_size; |
| 75 | |
| 76 | GEM_BUG_ON(!size); |
| 77 | |
| 78 | if (tiling == I915_TILING_NONE) |
| 79 | return size; |
| 80 | |
| 81 | GEM_BUG_ON(!stride); |
| 82 | |
| 83 | if (INTEL_GEN(i915) >= 4) { |
| 84 | stride *= i915_gem_tile_height(tiling); |
| 85 | GEM_BUG_ON(stride & 4095); |
| 86 | return roundup(size, stride); |
| 87 | } |
| 88 | |
| 89 | /* Previous chips need a power-of-two fence region when tiling */ |
| 90 | if (IS_GEN3(i915)) |
| 91 | ggtt_size = 1024*1024; |
| 92 | else |
| 93 | ggtt_size = 512*1024; |
| 94 | |
| 95 | while (ggtt_size < size) |
| 96 | ggtt_size <<= 1; |
| 97 | |
| 98 | return ggtt_size; |
| 99 | } |
| 100 | |
| 101 | /** |
| 102 | * i915_gem_fence_alignment - required global GTT alignment for a fence |
| 103 | * @i915: i915 device |
| 104 | * @size: object size |
| 105 | * @tiling: tiling mode |
| 106 | * @stride: tiling stride |
| 107 | * |
| 108 | * Return the required global GTT alignment for a fence (a view of a tiled |
| 109 | * object), taking into account potential fence register mapping. |
| 110 | */ |
| 111 | u32 i915_gem_fence_alignment(struct drm_i915_private *i915, u32 size, |
| 112 | unsigned int tiling, unsigned int stride) |
| 113 | { |
| 114 | GEM_BUG_ON(!size); |
| 115 | |
| 116 | /* |
| 117 | * Minimum alignment is 4k (GTT page size), but might be greater |
| 118 | * if a fence register is needed for the object. |
| 119 | */ |
| 120 | if (INTEL_GEN(i915) >= 4 || tiling == I915_TILING_NONE) |
| 121 | return 4096; |
| 122 | |
| 123 | /* |
| 124 | * Previous chips need to be aligned to the size of the smallest |
| 125 | * fence register that can contain the object. |
| 126 | */ |
| 127 | return i915_gem_fence_size(i915, size, tiling, stride); |
| 128 | } |
| 129 | |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 130 | /* Check pitch constriants for all chips & tiling formats */ |
Chris Wilson | a00b10c | 2010-09-24 21:15:47 +0100 | [diff] [blame] | 131 | static bool |
Tvrtko Ursulin | 118bb9f | 2016-11-16 08:55:36 +0000 | [diff] [blame] | 132 | i915_tiling_ok(struct drm_i915_private *dev_priv, |
| 133 | int stride, int size, int tiling_mode) |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 134 | { |
Chris Wilson | 0ee537a | 2011-03-06 09:03:16 +0000 | [diff] [blame] | 135 | int tile_width; |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 136 | |
| 137 | /* Linear is always fine */ |
| 138 | if (tiling_mode == I915_TILING_NONE) |
| 139 | return true; |
| 140 | |
Chris Wilson | deeb151 | 2016-08-05 10:14:22 +0100 | [diff] [blame] | 141 | if (tiling_mode > I915_TILING_LAST) |
| 142 | return false; |
| 143 | |
Tvrtko Ursulin | 5db9401 | 2016-10-13 11:03:10 +0100 | [diff] [blame] | 144 | if (IS_GEN2(dev_priv) || |
Tvrtko Ursulin | 50a0bc9 | 2016-10-13 11:02:58 +0100 | [diff] [blame] | 145 | (tiling_mode == I915_TILING_Y && HAS_128_BYTE_Y_TILING(dev_priv))) |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 146 | tile_width = 128; |
| 147 | else |
| 148 | tile_width = 512; |
| 149 | |
Daniel Vetter | 8d7773a | 2009-03-29 14:09:41 +0200 | [diff] [blame] | 150 | /* check maximum stride & object size */ |
Ville Syrjälä | 3a06247 | 2013-04-09 11:45:05 +0300 | [diff] [blame] | 151 | /* i965+ stores the end address of the gtt mapping in the fence |
| 152 | * reg, so dont bother to check the size */ |
Tvrtko Ursulin | 118bb9f | 2016-11-16 08:55:36 +0000 | [diff] [blame] | 153 | if (INTEL_GEN(dev_priv) >= 7) { |
Ville Syrjälä | 3a06247 | 2013-04-09 11:45:05 +0300 | [diff] [blame] | 154 | if (stride / 128 > GEN7_FENCE_MAX_PITCH_VAL) |
| 155 | return false; |
Tvrtko Ursulin | 118bb9f | 2016-11-16 08:55:36 +0000 | [diff] [blame] | 156 | } else if (INTEL_GEN(dev_priv) >= 4) { |
Daniel Vetter | 8d7773a | 2009-03-29 14:09:41 +0200 | [diff] [blame] | 157 | if (stride / 128 > I965_FENCE_MAX_PITCH_VAL) |
| 158 | return false; |
Chris Wilson | a6c45cf | 2010-09-17 00:32:17 +0100 | [diff] [blame] | 159 | } else { |
Daniel Vetter | c36a2a6 | 2010-04-17 15:12:03 +0200 | [diff] [blame] | 160 | if (stride > 8192) |
Daniel Vetter | 8d7773a | 2009-03-29 14:09:41 +0200 | [diff] [blame] | 161 | return false; |
Eric Anholt | e76a16d | 2009-05-26 17:44:56 -0700 | [diff] [blame] | 162 | |
Tvrtko Ursulin | 5db9401 | 2016-10-13 11:03:10 +0100 | [diff] [blame] | 163 | if (IS_GEN3(dev_priv)) { |
Daniel Vetter | c36a2a6 | 2010-04-17 15:12:03 +0200 | [diff] [blame] | 164 | if (size > I830_FENCE_MAX_SIZE_VAL << 20) |
| 165 | return false; |
| 166 | } else { |
| 167 | if (size > I830_FENCE_MAX_SIZE_VAL << 19) |
| 168 | return false; |
| 169 | } |
Daniel Vetter | 8d7773a | 2009-03-29 14:09:41 +0200 | [diff] [blame] | 170 | } |
| 171 | |
Ville Syrjälä | fe48d8d | 2013-04-09 20:09:13 +0300 | [diff] [blame] | 172 | if (stride < tile_width) |
| 173 | return false; |
| 174 | |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 175 | /* 965+ just needs multiples of tile width */ |
Tvrtko Ursulin | 118bb9f | 2016-11-16 08:55:36 +0000 | [diff] [blame] | 176 | if (INTEL_GEN(dev_priv) >= 4) { |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 177 | if (stride & (tile_width - 1)) |
| 178 | return false; |
| 179 | return true; |
| 180 | } |
| 181 | |
| 182 | /* Pre-965 needs power of two tile widths */ |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 183 | if (stride & (stride - 1)) |
| 184 | return false; |
| 185 | |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 186 | return true; |
| 187 | } |
| 188 | |
Chris Wilson | 5b30694 | 2017-01-09 16:16:09 +0000 | [diff] [blame] | 189 | static bool i915_vma_fence_prepare(struct i915_vma *vma, |
| 190 | int tiling_mode, unsigned int stride) |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 191 | { |
Chris Wilson | 944397f | 2017-01-09 16:16:11 +0000 | [diff] [blame] | 192 | struct drm_i915_private *i915 = vma->vm->i915; |
| 193 | u32 size, alignment; |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 194 | |
| 195 | if (!i915_vma_is_map_and_fenceable(vma)) |
| 196 | return true; |
| 197 | |
Chris Wilson | 91d4e0aa | 2017-01-09 16:16:13 +0000 | [diff] [blame] | 198 | size = i915_gem_fence_size(i915, vma->size, tiling_mode, stride); |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 199 | if (vma->node.size < size) |
| 200 | return false; |
| 201 | |
Chris Wilson | 91d4e0aa | 2017-01-09 16:16:13 +0000 | [diff] [blame] | 202 | alignment = i915_gem_fence_alignment(i915, vma->size, tiling_mode, stride); |
Chris Wilson | 944397f | 2017-01-09 16:16:11 +0000 | [diff] [blame] | 203 | if (vma->node.start & (alignment - 1)) |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 204 | return false; |
| 205 | |
| 206 | return true; |
| 207 | } |
| 208 | |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 209 | /* Make the current GTT allocation valid for the change in tiling. */ |
| 210 | static int |
Chris Wilson | 5b30694 | 2017-01-09 16:16:09 +0000 | [diff] [blame] | 211 | i915_gem_object_fence_prepare(struct drm_i915_gem_object *obj, |
| 212 | int tiling_mode, unsigned int stride) |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 213 | { |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 214 | struct i915_vma *vma; |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 215 | int ret; |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 216 | |
| 217 | if (tiling_mode == I915_TILING_NONE) |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 218 | return 0; |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 219 | |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 220 | list_for_each_entry(vma, &obj->vma_list, obj_link) { |
Chris Wilson | 944397f | 2017-01-09 16:16:11 +0000 | [diff] [blame] | 221 | if (!i915_vma_is_ggtt(vma)) |
| 222 | break; |
| 223 | |
Chris Wilson | 5b30694 | 2017-01-09 16:16:09 +0000 | [diff] [blame] | 224 | if (i915_vma_fence_prepare(vma, tiling_mode, stride)) |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 225 | continue; |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 226 | |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 227 | ret = i915_vma_unbind(vma); |
| 228 | if (ret) |
| 229 | return ret; |
Chris Wilson | df15315 | 2010-11-15 05:25:58 +0000 | [diff] [blame] | 230 | } |
| 231 | |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 232 | return 0; |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 233 | } |
| 234 | |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 235 | /** |
Chris Wilson | 111dbca | 2017-01-10 12:10:44 +0000 | [diff] [blame^] | 236 | * i915_gem_set_tiling_ioctl - IOCTL handler to set tiling mode |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 237 | * @dev: DRM device |
| 238 | * @data: data pointer for the ioctl |
| 239 | * @file: DRM file for the ioctl call |
| 240 | * |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 241 | * Sets the tiling mode of an object, returning the required swizzling of |
| 242 | * bit 6 of addresses in the object. |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 243 | * |
| 244 | * Called by the user via ioctl. |
| 245 | * |
| 246 | * Returns: |
| 247 | * Zero on success, negative errno on failure. |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 248 | */ |
| 249 | int |
Chris Wilson | 111dbca | 2017-01-10 12:10:44 +0000 | [diff] [blame^] | 250 | i915_gem_set_tiling_ioctl(struct drm_device *dev, void *data, |
| 251 | struct drm_file *file) |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 252 | { |
| 253 | struct drm_i915_gem_set_tiling *args = data; |
Chris Wilson | fac5e23 | 2016-07-04 11:34:36 +0100 | [diff] [blame] | 254 | struct drm_i915_private *dev_priv = to_i915(dev); |
Chris Wilson | 05394f3 | 2010-11-08 19:18:58 +0000 | [diff] [blame] | 255 | struct drm_i915_gem_object *obj; |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 256 | int err = 0; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 257 | |
Chris Wilson | 3e510a8 | 2016-08-05 10:14:23 +0100 | [diff] [blame] | 258 | /* Make sure we don't cross-contaminate obj->tiling_and_stride */ |
| 259 | BUILD_BUG_ON(I915_TILING_LAST & STRIDE_MASK); |
| 260 | |
Chris Wilson | 03ac064 | 2016-07-20 13:31:51 +0100 | [diff] [blame] | 261 | obj = i915_gem_object_lookup(file, args->handle); |
| 262 | if (!obj) |
Chris Wilson | bf79cb9 | 2010-08-04 14:19:46 +0100 | [diff] [blame] | 263 | return -ENOENT; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 264 | |
Tvrtko Ursulin | 118bb9f | 2016-11-16 08:55:36 +0000 | [diff] [blame] | 265 | if (!i915_tiling_ok(dev_priv, |
Chris Wilson | 05394f3 | 2010-11-08 19:18:58 +0000 | [diff] [blame] | 266 | args->stride, obj->base.size, args->tiling_mode)) { |
Chris Wilson | f0cd518 | 2016-10-28 13:58:43 +0100 | [diff] [blame] | 267 | i915_gem_object_put(obj); |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 268 | return -EINVAL; |
Chris Wilson | 72daad4 | 2009-01-30 21:10:22 +0000 | [diff] [blame] | 269 | } |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 270 | |
Chris Wilson | 6c31a61 | 2015-02-12 07:53:18 +0000 | [diff] [blame] | 271 | mutex_lock(&dev->struct_mutex); |
Chris Wilson | 1f30a61 | 2015-04-15 16:39:59 +0100 | [diff] [blame] | 272 | if (obj->pin_display || obj->framebuffer_references) { |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 273 | err = -EBUSY; |
Chris Wilson | 6c31a61 | 2015-02-12 07:53:18 +0000 | [diff] [blame] | 274 | goto err; |
Daniel Vetter | 31770bd | 2010-04-23 23:01:01 +0200 | [diff] [blame] | 275 | } |
| 276 | |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 277 | if (args->tiling_mode == I915_TILING_NONE) { |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 278 | args->swizzle_mode = I915_BIT_6_SWIZZLE_NONE; |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 279 | args->stride = 0; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 280 | } else { |
| 281 | if (args->tiling_mode == I915_TILING_X) |
| 282 | args->swizzle_mode = dev_priv->mm.bit_6_swizzle_x; |
| 283 | else |
| 284 | args->swizzle_mode = dev_priv->mm.bit_6_swizzle_y; |
Eric Anholt | 280b713 | 2009-03-12 16:56:27 -0700 | [diff] [blame] | 285 | |
| 286 | /* Hide bit 17 swizzling from the user. This prevents old Mesa |
| 287 | * from aborting the application on sw fallbacks to bit 17, |
| 288 | * and we use the pread/pwrite bit17 paths to swizzle for it. |
| 289 | * If there was a user that was relying on the swizzle |
| 290 | * information for drm_intel_bo_map()ed reads/writes this would |
| 291 | * break it, but we don't have any of those. |
| 292 | */ |
| 293 | if (args->swizzle_mode == I915_BIT_6_SWIZZLE_9_17) |
| 294 | args->swizzle_mode = I915_BIT_6_SWIZZLE_9; |
| 295 | if (args->swizzle_mode == I915_BIT_6_SWIZZLE_9_10_17) |
| 296 | args->swizzle_mode = I915_BIT_6_SWIZZLE_9_10; |
| 297 | |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 298 | /* If we can't handle the swizzling, make it untiled. */ |
| 299 | if (args->swizzle_mode == I915_BIT_6_SWIZZLE_UNKNOWN) { |
| 300 | args->tiling_mode = I915_TILING_NONE; |
| 301 | args->swizzle_mode = I915_BIT_6_SWIZZLE_NONE; |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 302 | args->stride = 0; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 303 | } |
| 304 | } |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 305 | |
Chris Wilson | 3e510a8 | 2016-08-05 10:14:23 +0100 | [diff] [blame] | 306 | if (args->tiling_mode != i915_gem_object_get_tiling(obj) || |
| 307 | args->stride != i915_gem_object_get_stride(obj)) { |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 308 | /* We need to rebind the object if its current allocation |
| 309 | * no longer meets the alignment restrictions for its new |
| 310 | * tiling mode. Otherwise we can just leave it alone, but |
Chris Wilson | 1869b62 | 2012-04-21 16:23:24 +0100 | [diff] [blame] | 311 | * need to ensure that any fence register is updated before |
| 312 | * the next fenced (either through the GTT or by the BLT unit |
| 313 | * on older GPUs) access. |
Chris Wilson | 5d82e3e | 2012-04-21 16:23:23 +0100 | [diff] [blame] | 314 | * |
| 315 | * After updating the tiling parameters, we then flag whether |
| 316 | * we need to update an associated fence register. Note this |
| 317 | * has to also include the unfenced register the GPU uses |
| 318 | * whilst executing a fenced command for an untiled object. |
Jesse Barnes | 0f973f2 | 2009-01-26 17:10:45 -0800 | [diff] [blame] | 319 | */ |
Chris Wilson | 467cffb | 2011-03-07 10:42:03 +0000 | [diff] [blame] | 320 | |
Chris Wilson | 5b30694 | 2017-01-09 16:16:09 +0000 | [diff] [blame] | 321 | err = i915_gem_object_fence_prepare(obj, |
| 322 | args->tiling_mode, |
| 323 | args->stride); |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 324 | if (!err) { |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 325 | struct i915_vma *vma; |
| 326 | |
Chris Wilson | 1233e2d | 2016-10-28 13:58:37 +0100 | [diff] [blame] | 327 | mutex_lock(&obj->mm.lock); |
Chris Wilson | a4f5ea6 | 2016-10-28 13:58:35 +0100 | [diff] [blame] | 328 | if (obj->mm.pages && |
| 329 | obj->mm.madv == I915_MADV_WILLNEED && |
Daniel Vetter | 656bfa3 | 2014-11-20 09:26:30 +0100 | [diff] [blame] | 330 | dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) { |
Chris Wilson | bc0629a | 2016-11-01 10:03:17 +0000 | [diff] [blame] | 331 | if (args->tiling_mode == I915_TILING_NONE) { |
| 332 | GEM_BUG_ON(!obj->mm.quirked); |
Chris Wilson | a4f5ea6 | 2016-10-28 13:58:35 +0100 | [diff] [blame] | 333 | __i915_gem_object_unpin_pages(obj); |
Chris Wilson | bc0629a | 2016-11-01 10:03:17 +0000 | [diff] [blame] | 334 | obj->mm.quirked = false; |
| 335 | } |
| 336 | if (!i915_gem_object_is_tiled(obj)) { |
Chris Wilson | 2c3a3f4 | 2016-11-04 10:30:01 +0000 | [diff] [blame] | 337 | GEM_BUG_ON(!obj->mm.quirked); |
Chris Wilson | a4f5ea6 | 2016-10-28 13:58:35 +0100 | [diff] [blame] | 338 | __i915_gem_object_pin_pages(obj); |
Chris Wilson | bc0629a | 2016-11-01 10:03:17 +0000 | [diff] [blame] | 339 | obj->mm.quirked = true; |
| 340 | } |
Daniel Vetter | 656bfa3 | 2014-11-20 09:26:30 +0100 | [diff] [blame] | 341 | } |
Chris Wilson | 1233e2d | 2016-10-28 13:58:37 +0100 | [diff] [blame] | 342 | mutex_unlock(&obj->mm.lock); |
Daniel Vetter | 656bfa3 | 2014-11-20 09:26:30 +0100 | [diff] [blame] | 343 | |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 344 | list_for_each_entry(vma, &obj->vma_list, obj_link) { |
Chris Wilson | 944397f | 2017-01-09 16:16:11 +0000 | [diff] [blame] | 345 | if (!i915_vma_is_ggtt(vma)) |
| 346 | break; |
Chris Wilson | 5d82e3e | 2012-04-21 16:23:23 +0100 | [diff] [blame] | 347 | |
Chris Wilson | 91d4e0aa | 2017-01-09 16:16:13 +0000 | [diff] [blame] | 348 | vma->fence_size = i915_gem_fence_size(dev_priv, vma->size, |
| 349 | args->tiling_mode, |
| 350 | args->stride); |
| 351 | vma->fence_alignment = i915_gem_fence_alignment(dev_priv, vma->size, |
| 352 | args->tiling_mode, |
| 353 | args->stride); |
Chris Wilson | 944397f | 2017-01-09 16:16:11 +0000 | [diff] [blame] | 354 | |
| 355 | if (vma->fence) |
| 356 | vma->fence->dirty = true; |
Chris Wilson | 49ef529 | 2016-08-18 17:17:00 +0100 | [diff] [blame] | 357 | } |
Chris Wilson | 3e510a8 | 2016-08-05 10:14:23 +0100 | [diff] [blame] | 358 | obj->tiling_and_stride = |
| 359 | args->stride | args->tiling_mode; |
Chris Wilson | 1869b62 | 2012-04-21 16:23:24 +0100 | [diff] [blame] | 360 | |
| 361 | /* Force the fence to be reacquired for GTT access */ |
| 362 | i915_gem_release_mmap(obj); |
Chris Wilson | 467cffb | 2011-03-07 10:42:03 +0000 | [diff] [blame] | 363 | } |
Chris Wilson | 52dc7d3 | 2009-06-06 09:46:01 +0100 | [diff] [blame] | 364 | } |
Chris Wilson | 467cffb | 2011-03-07 10:42:03 +0000 | [diff] [blame] | 365 | /* we have to maintain this existing ABI... */ |
Chris Wilson | 3e510a8 | 2016-08-05 10:14:23 +0100 | [diff] [blame] | 366 | args->stride = i915_gem_object_get_stride(obj); |
| 367 | args->tiling_mode = i915_gem_object_get_tiling(obj); |
Chris Wilson | e9b73c6 | 2012-12-03 21:03:14 +0000 | [diff] [blame] | 368 | |
| 369 | /* Try to preallocate memory required to save swizzling on put-pages */ |
| 370 | if (i915_gem_object_needs_bit17_swizzle(obj)) { |
| 371 | if (obj->bit_17 == NULL) { |
Daniel Vetter | a1e2265 | 2013-09-21 00:35:38 +0200 | [diff] [blame] | 372 | obj->bit_17 = kcalloc(BITS_TO_LONGS(obj->base.size >> PAGE_SHIFT), |
Chris Wilson | e9b73c6 | 2012-12-03 21:03:14 +0000 | [diff] [blame] | 373 | sizeof(long), GFP_KERNEL); |
| 374 | } |
| 375 | } else { |
| 376 | kfree(obj->bit_17); |
| 377 | obj->bit_17 = NULL; |
| 378 | } |
| 379 | |
Chris Wilson | 6c31a61 | 2015-02-12 07:53:18 +0000 | [diff] [blame] | 380 | err: |
Chris Wilson | f8c417c | 2016-07-20 13:31:53 +0100 | [diff] [blame] | 381 | i915_gem_object_put(obj); |
Chris Wilson | d687310 | 2009-02-08 19:07:51 +0000 | [diff] [blame] | 382 | mutex_unlock(&dev->struct_mutex); |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 383 | |
Chris Wilson | f23eda8 | 2016-08-15 10:48:53 +0100 | [diff] [blame] | 384 | return err; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 385 | } |
| 386 | |
| 387 | /** |
Chris Wilson | 111dbca | 2017-01-10 12:10:44 +0000 | [diff] [blame^] | 388 | * i915_gem_get_tiling_ioctl - IOCTL handler to get tiling mode |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 389 | * @dev: DRM device |
| 390 | * @data: data pointer for the ioctl |
| 391 | * @file: DRM file for the ioctl call |
| 392 | * |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 393 | * Returns the current tiling mode and required bit 6 swizzling for the object. |
Daniel Vetter | 3271dca | 2015-07-24 17:40:15 +0200 | [diff] [blame] | 394 | * |
| 395 | * Called by the user via ioctl. |
| 396 | * |
| 397 | * Returns: |
| 398 | * Zero on success, negative errno on failure. |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 399 | */ |
| 400 | int |
Chris Wilson | 111dbca | 2017-01-10 12:10:44 +0000 | [diff] [blame^] | 401 | i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data, |
| 402 | struct drm_file *file) |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 403 | { |
| 404 | struct drm_i915_gem_get_tiling *args = data; |
Chris Wilson | fac5e23 | 2016-07-04 11:34:36 +0100 | [diff] [blame] | 405 | struct drm_i915_private *dev_priv = to_i915(dev); |
Chris Wilson | 05394f3 | 2010-11-08 19:18:58 +0000 | [diff] [blame] | 406 | struct drm_i915_gem_object *obj; |
Chris Wilson | fbbd37b | 2016-10-28 13:58:42 +0100 | [diff] [blame] | 407 | int err = -ENOENT; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 408 | |
Chris Wilson | fbbd37b | 2016-10-28 13:58:42 +0100 | [diff] [blame] | 409 | rcu_read_lock(); |
| 410 | obj = i915_gem_object_lookup_rcu(file, args->handle); |
| 411 | if (obj) { |
| 412 | args->tiling_mode = |
| 413 | READ_ONCE(obj->tiling_and_stride) & TILING_MASK; |
| 414 | err = 0; |
| 415 | } |
| 416 | rcu_read_unlock(); |
| 417 | if (unlikely(err)) |
| 418 | return err; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 419 | |
Chris Wilson | 9ad3676 | 2016-08-05 10:14:21 +0100 | [diff] [blame] | 420 | switch (args->tiling_mode) { |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 421 | case I915_TILING_X: |
| 422 | args->swizzle_mode = dev_priv->mm.bit_6_swizzle_x; |
| 423 | break; |
| 424 | case I915_TILING_Y: |
| 425 | args->swizzle_mode = dev_priv->mm.bit_6_swizzle_y; |
| 426 | break; |
Chris Wilson | fbbd37b | 2016-10-28 13:58:42 +0100 | [diff] [blame] | 427 | default: |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 428 | case I915_TILING_NONE: |
| 429 | args->swizzle_mode = I915_BIT_6_SWIZZLE_NONE; |
| 430 | break; |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 431 | } |
| 432 | |
Eric Anholt | 280b713 | 2009-03-12 16:56:27 -0700 | [diff] [blame] | 433 | /* Hide bit 17 from the user -- see comment in i915_gem_set_tiling */ |
Chris Wilson | 5eb3e5a | 2015-06-28 09:19:26 +0100 | [diff] [blame] | 434 | if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) |
| 435 | args->phys_swizzle_mode = I915_BIT_6_SWIZZLE_UNKNOWN; |
| 436 | else |
| 437 | args->phys_swizzle_mode = args->swizzle_mode; |
Eric Anholt | 280b713 | 2009-03-12 16:56:27 -0700 | [diff] [blame] | 438 | if (args->swizzle_mode == I915_BIT_6_SWIZZLE_9_17) |
| 439 | args->swizzle_mode = I915_BIT_6_SWIZZLE_9; |
| 440 | if (args->swizzle_mode == I915_BIT_6_SWIZZLE_9_10_17) |
| 441 | args->swizzle_mode = I915_BIT_6_SWIZZLE_9_10; |
| 442 | |
Eric Anholt | 673a394 | 2008-07-30 12:06:12 -0700 | [diff] [blame] | 443 | return 0; |
| 444 | } |