blob: 208768922030a7ff792fa958d772cb72f9ebb4cb [file] [log] [blame]
Rob Clark16ea9752013-01-08 15:04:28 -06001/*
2 * Copyright (C) 2012 Texas Instruments
3 * Author: Rob Clark <robdclark@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
16 */
17
Rob Clarka464d612013-08-07 13:41:20 -040018#include "drm_flip_work.h"
Daniel Vetter3cb9ae42014-10-29 10:03:57 +010019#include <drm/drm_plane_helper.h>
Jyri Sarha305198d2016-04-07 15:05:16 +030020#include <drm/drm_atomic_helper.h>
Rob Clark16ea9752013-01-08 15:04:28 -060021
22#include "tilcdc_drv.h"
23#include "tilcdc_regs.h"
24
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020025#define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
26
Rob Clark16ea9752013-01-08 15:04:28 -060027struct tilcdc_crtc {
28 struct drm_crtc base;
29
Jyri Sarha47f571c2016-04-07 15:04:18 +030030 struct drm_plane primary;
Rob Clark16ea9752013-01-08 15:04:28 -060031 const struct tilcdc_panel_info *info;
Rob Clark16ea9752013-01-08 15:04:28 -060032 struct drm_pending_vblank_event *event;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +030033 bool enabled;
Rob Clark16ea9752013-01-08 15:04:28 -060034 wait_queue_head_t frame_done_wq;
35 bool frame_done;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020036 spinlock_t irq_lock;
37
38 ktime_t last_vblank;
Rob Clark16ea9752013-01-08 15:04:28 -060039
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030040 struct drm_framebuffer *curr_fb;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020041 struct drm_framebuffer *next_fb;
Rob Clark16ea9752013-01-08 15:04:28 -060042
43 /* for deferred fb unref's: */
Rob Clarka464d612013-08-07 13:41:20 -040044 struct drm_flip_work unref_work;
Jyri Sarha103cd8b2015-02-10 14:13:23 +020045
46 /* Only set if an external encoder is connected */
47 bool simulate_vesa_sync;
Jyri Sarha5895d082016-01-08 14:33:09 +020048
49 int sync_lost_count;
50 bool frame_intact;
Rob Clark16ea9752013-01-08 15:04:28 -060051};
52#define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
53
Rob Clarka464d612013-08-07 13:41:20 -040054static void unref_worker(struct drm_flip_work *work, void *val)
Rob Clark16ea9752013-01-08 15:04:28 -060055{
Darren Etheridgef7b45752013-06-21 13:52:26 -050056 struct tilcdc_crtc *tilcdc_crtc =
Rob Clarka464d612013-08-07 13:41:20 -040057 container_of(work, struct tilcdc_crtc, unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -060058 struct drm_device *dev = tilcdc_crtc->base.dev;
Rob Clark16ea9752013-01-08 15:04:28 -060059
60 mutex_lock(&dev->mode_config.mutex);
Rob Clarka464d612013-08-07 13:41:20 -040061 drm_framebuffer_unreference(val);
Rob Clark16ea9752013-01-08 15:04:28 -060062 mutex_unlock(&dev->mode_config.mutex);
63}
64
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030065static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
Rob Clark16ea9752013-01-08 15:04:28 -060066{
67 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
68 struct drm_device *dev = crtc->dev;
Rob Clark16ea9752013-01-08 15:04:28 -060069 struct drm_gem_cma_object *gem;
70 unsigned int depth, bpp;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030071 dma_addr_t start, end;
Jyri Sarha7eb9f062016-08-26 15:10:14 +030072 u64 dma_base_and_ceiling;
Rob Clark16ea9752013-01-08 15:04:28 -060073
74 drm_fb_get_bpp_depth(fb->pixel_format, &depth, &bpp);
75 gem = drm_fb_cma_get_gem_obj(fb, 0);
76
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030077 start = gem->paddr + fb->offsets[0] +
78 crtc->y * fb->pitches[0] +
79 crtc->x * bpp / 8;
Rob Clark16ea9752013-01-08 15:04:28 -060080
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030081 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
Rob Clark16ea9752013-01-08 15:04:28 -060082
Jyri Sarha7eb9f062016-08-26 15:10:14 +030083 /* Write LCDC_DMA_FB_BASE_ADDR_0_REG and LCDC_DMA_FB_CEILING_ADDR_0_REG
84 * with a single insruction, if available. This should make it more
85 * unlikely that LCDC would fetch the DMA addresses in the middle of
86 * an update.
87 */
88 dma_base_and_ceiling = (u64)(end - 1) << 32 | start;
89 tilcdc_write64(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_base_and_ceiling);
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +030090
91 if (tilcdc_crtc->curr_fb)
92 drm_flip_work_queue(&tilcdc_crtc->unref_work,
93 tilcdc_crtc->curr_fb);
94
95 tilcdc_crtc->curr_fb = fb;
Rob Clark16ea9752013-01-08 15:04:28 -060096}
97
Jyri Sarhaafaf8332016-06-21 16:00:44 +030098static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
99{
100 struct tilcdc_drm_private *priv = dev->dev_private;
101
102 tilcdc_clear_irqstatus(dev, 0xffffffff);
103
104 if (priv->rev == 1) {
105 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
106 LCDC_V1_UNDERFLOW_INT_ENA);
Karl Beldan8d6c3f72016-08-23 12:57:00 +0000107 tilcdc_set(dev, LCDC_DMA_CTRL_REG,
108 LCDC_V1_END_OF_FRAME_INT_ENA);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300109 } else {
110 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
111 LCDC_V2_UNDERFLOW_INT_ENA |
112 LCDC_V2_END_OF_FRAME0_INT_ENA |
113 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
114 }
115}
116
117static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
118{
119 struct tilcdc_drm_private *priv = dev->dev_private;
120
121 /* disable irqs that we might have enabled: */
122 if (priv->rev == 1) {
123 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
124 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
125 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
126 LCDC_V1_END_OF_FRAME_INT_ENA);
127 } else {
128 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
129 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
130 LCDC_V2_END_OF_FRAME0_INT_ENA |
131 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
132 }
133}
134
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300135static void reset(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600136{
137 struct drm_device *dev = crtc->dev;
138 struct tilcdc_drm_private *priv = dev->dev_private;
139
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300140 if (priv->rev != 2)
141 return;
142
143 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
144 usleep_range(250, 1000);
145 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
146}
147
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300148static void tilcdc_crtc_enable(struct drm_crtc *crtc)
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300149{
150 struct drm_device *dev = crtc->dev;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300151 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
152
153 if (tilcdc_crtc->enabled)
154 return;
155
156 pm_runtime_get_sync(dev->dev);
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300157
158 reset(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600159
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300160 tilcdc_crtc_enable_irqs(dev);
161
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300162 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
Rob Clark16ea9752013-01-08 15:04:28 -0600163 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_PALETTE_LOAD_MODE(DATA_ONLY));
164 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300165
166 drm_crtc_vblank_on(crtc);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300167
168 tilcdc_crtc->enabled = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600169}
170
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300171void tilcdc_crtc_disable(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600172{
Jyri Sarha2d5be882016-04-07 20:20:23 +0300173 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600174 struct drm_device *dev = crtc->dev;
Jyri Sarha2d5be882016-04-07 20:20:23 +0300175 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600176
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300177 if (!tilcdc_crtc->enabled)
178 return;
179
Jyri Sarha2d5be882016-04-07 20:20:23 +0300180 tilcdc_crtc->frame_done = false;
Rob Clark16ea9752013-01-08 15:04:28 -0600181 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha2d5be882016-04-07 20:20:23 +0300182
183 /*
184 * if necessary wait for framedone irq which will still come
185 * before putting things to sleep..
186 */
187 if (priv->rev == 2) {
188 int ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
189 tilcdc_crtc->frame_done,
Jyri Sarha437c7d92016-06-16 16:19:17 +0300190 msecs_to_jiffies(500));
Jyri Sarha2d5be882016-04-07 20:20:23 +0300191 if (ret == 0)
192 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
193 __func__);
194 }
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300195
196 drm_crtc_vblank_off(crtc);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300197
198 tilcdc_crtc_disable_irqs(dev);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300199
200 pm_runtime_put_sync(dev->dev);
201
202 if (tilcdc_crtc->next_fb) {
203 drm_flip_work_queue(&tilcdc_crtc->unref_work,
204 tilcdc_crtc->next_fb);
205 tilcdc_crtc->next_fb = NULL;
206 }
207
208 if (tilcdc_crtc->curr_fb) {
209 drm_flip_work_queue(&tilcdc_crtc->unref_work,
210 tilcdc_crtc->curr_fb);
211 tilcdc_crtc->curr_fb = NULL;
212 }
213
214 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
215 tilcdc_crtc->last_vblank = ktime_set(0, 0);
216
217 tilcdc_crtc->enabled = false;
218}
219
220static bool tilcdc_crtc_is_on(struct drm_crtc *crtc)
221{
222 return crtc->state && crtc->state->enable && crtc->state->active;
Rob Clark16ea9752013-01-08 15:04:28 -0600223}
224
225static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
226{
227 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
228
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300229 tilcdc_crtc_disable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600230
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300231 of_node_put(crtc->port);
Rob Clark16ea9752013-01-08 15:04:28 -0600232 drm_crtc_cleanup(crtc);
Rob Clarka464d612013-08-07 13:41:20 -0400233 drm_flip_work_cleanup(&tilcdc_crtc->unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -0600234}
235
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300236int tilcdc_crtc_update_fb(struct drm_crtc *crtc,
Rob Clark16ea9752013-01-08 15:04:28 -0600237 struct drm_framebuffer *fb,
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300238 struct drm_pending_vblank_event *event)
Rob Clark16ea9752013-01-08 15:04:28 -0600239{
240 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
241 struct drm_device *dev = crtc->dev;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300242 unsigned long flags;
Tomi Valkeinen6f206e92014-02-07 17:37:07 +0000243
Rob Clark16ea9752013-01-08 15:04:28 -0600244 if (tilcdc_crtc->event) {
245 dev_err(dev->dev, "already pending page flip!\n");
246 return -EBUSY;
247 }
248
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300249 drm_framebuffer_reference(fb);
250
Matt Roperf4510a22014-04-01 15:22:40 -0700251 crtc->primary->fb = fb;
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300252
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200253 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300254
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300255 if (crtc->hwmode.vrefresh && ktime_to_ns(tilcdc_crtc->last_vblank)) {
256 ktime_t next_vblank;
257 s64 tdiff;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300258
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300259 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
260 1000000 / crtc->hwmode.vrefresh);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200261
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300262 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
263
264 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
265 tilcdc_crtc->next_fb = fb;
266 }
267
268 if (tilcdc_crtc->next_fb != fb)
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200269 set_scanout(crtc, fb);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200270
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300271 tilcdc_crtc->event = event;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200272
273 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600274
275 return 0;
276}
277
Rob Clark16ea9752013-01-08 15:04:28 -0600278static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
279 const struct drm_display_mode *mode,
280 struct drm_display_mode *adjusted_mode)
281{
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200282 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
283
284 if (!tilcdc_crtc->simulate_vesa_sync)
285 return true;
286
287 /*
288 * tilcdc does not generate VESA-compliant sync but aligns
289 * VS on the second edge of HS instead of first edge.
290 * We use adjusted_mode, to fixup sync by aligning both rising
291 * edges and add HSKEW offset to fix the sync.
292 */
293 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
294 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
295
296 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
297 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
298 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
299 } else {
300 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
301 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
302 }
303
Rob Clark16ea9752013-01-08 15:04:28 -0600304 return true;
305}
306
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300307static void tilcdc_crtc_mode_set_nofb(struct drm_crtc *crtc)
308{
309 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
310 struct drm_device *dev = crtc->dev;
311 struct tilcdc_drm_private *priv = dev->dev_private;
312 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
313 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
314 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
315 struct drm_framebuffer *fb = crtc->primary->state->fb;
316
317 if (WARN_ON(!info))
318 return;
319
320 if (WARN_ON(!fb))
321 return;
322
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300323 /* Configure the Burst Size and fifo threshold of DMA: */
324 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
325 switch (info->dma_burst_sz) {
326 case 1:
327 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
328 break;
329 case 2:
330 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
331 break;
332 case 4:
333 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
334 break;
335 case 8:
336 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
337 break;
338 case 16:
339 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
340 break;
341 default:
342 dev_err(dev->dev, "invalid burst size\n");
343 return;
344 }
345 reg |= (info->fifo_th << 8);
346 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
347
348 /* Configure timings: */
349 hbp = mode->htotal - mode->hsync_end;
350 hfp = mode->hsync_start - mode->hdisplay;
351 hsw = mode->hsync_end - mode->hsync_start;
352 vbp = mode->vtotal - mode->vsync_end;
353 vfp = mode->vsync_start - mode->vdisplay;
354 vsw = mode->vsync_end - mode->vsync_start;
355
356 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
357 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
358
359 /* Set AC Bias Period and Number of Transitions per Interrupt: */
360 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
361 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
362 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
363
364 /*
365 * subtract one from hfp, hbp, hsw because the hardware uses
366 * a value of 0 as 1
367 */
368 if (priv->rev == 2) {
369 /* clear bits we're going to set */
370 reg &= ~0x78000033;
371 reg |= ((hfp-1) & 0x300) >> 8;
372 reg |= ((hbp-1) & 0x300) >> 4;
373 reg |= ((hsw-1) & 0x3c0) << 21;
374 }
375 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
376
377 reg = (((mode->hdisplay >> 4) - 1) << 4) |
378 (((hbp-1) & 0xff) << 24) |
379 (((hfp-1) & 0xff) << 16) |
380 (((hsw-1) & 0x3f) << 10);
381 if (priv->rev == 2)
382 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
383 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
384
385 reg = ((mode->vdisplay - 1) & 0x3ff) |
386 ((vbp & 0xff) << 24) |
387 ((vfp & 0xff) << 16) |
388 (((vsw-1) & 0x3f) << 10);
389 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
390
391 /*
392 * be sure to set Bit 10 for the V2 LCDC controller,
393 * otherwise limited to 1024 pixels width, stopping
394 * 1920x1080 being supported.
395 */
396 if (priv->rev == 2) {
397 if ((mode->vdisplay - 1) & 0x400) {
398 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
399 LCDC_LPP_B10);
400 } else {
401 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
402 LCDC_LPP_B10);
403 }
404 }
405
406 /* Configure display type: */
407 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
408 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
409 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
410 0x000ff000 /* Palette Loading Delay bits */);
411 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
412 if (info->tft_alt_mode)
413 reg |= LCDC_TFT_ALT_ENABLE;
414 if (priv->rev == 2) {
415 unsigned int depth, bpp;
416
417 drm_fb_get_bpp_depth(fb->pixel_format, &depth, &bpp);
418 switch (bpp) {
419 case 16:
420 break;
421 case 32:
422 reg |= LCDC_V2_TFT_24BPP_UNPACK;
423 /* fallthrough */
424 case 24:
425 reg |= LCDC_V2_TFT_24BPP_MODE;
426 break;
427 default:
428 dev_err(dev->dev, "invalid pixel format\n");
429 return;
430 }
431 }
432 reg |= info->fdd < 12;
433 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
434
435 if (info->invert_pxl_clk)
436 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
437 else
438 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
439
440 if (info->sync_ctrl)
441 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
442 else
443 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
444
445 if (info->sync_edge)
446 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
447 else
448 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
449
450 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
451 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
452 else
453 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
454
455 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
456 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
457 else
458 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
459
460 if (info->raster_order)
461 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
462 else
463 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
464
465 drm_framebuffer_reference(fb);
466
467 set_scanout(crtc, fb);
468
469 tilcdc_crtc_update_clk(crtc);
470
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300471 crtc->hwmode = crtc->state->adjusted_mode;
472}
473
Jyri Sarhadb380c52016-04-07 15:10:23 +0300474static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
475 struct drm_crtc_state *state)
476{
477 struct drm_display_mode *mode = &state->mode;
478 int ret;
479
480 /* If we are not active we don't care */
481 if (!state->active)
482 return 0;
483
484 if (state->state->planes[0].ptr != crtc->primary ||
485 state->state->planes[0].state == NULL ||
486 state->state->planes[0].state->crtc != crtc) {
487 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
488 return -EINVAL;
489 }
490
491 ret = tilcdc_crtc_mode_valid(crtc, mode);
492 if (ret) {
493 dev_dbg(crtc->dev->dev, "Mode \"%s\" not valid", mode->name);
494 return -EINVAL;
495 }
496
497 return 0;
498}
499
Rob Clark16ea9752013-01-08 15:04:28 -0600500static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
Jyri Sarha305198d2016-04-07 15:05:16 +0300501 .destroy = tilcdc_crtc_destroy,
502 .set_config = drm_atomic_helper_set_config,
503 .page_flip = drm_atomic_helper_page_flip,
504 .reset = drm_atomic_helper_crtc_reset,
505 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
506 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
Rob Clark16ea9752013-01-08 15:04:28 -0600507};
508
509static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
Rob Clark16ea9752013-01-08 15:04:28 -0600510 .mode_fixup = tilcdc_crtc_mode_fixup,
Jyri Sarha305198d2016-04-07 15:05:16 +0300511 .enable = tilcdc_crtc_enable,
512 .disable = tilcdc_crtc_disable,
Jyri Sarhadb380c52016-04-07 15:10:23 +0300513 .atomic_check = tilcdc_crtc_atomic_check,
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300514 .mode_set_nofb = tilcdc_crtc_mode_set_nofb,
Rob Clark16ea9752013-01-08 15:04:28 -0600515};
516
517int tilcdc_crtc_max_width(struct drm_crtc *crtc)
518{
519 struct drm_device *dev = crtc->dev;
520 struct tilcdc_drm_private *priv = dev->dev_private;
521 int max_width = 0;
522
523 if (priv->rev == 1)
524 max_width = 1024;
525 else if (priv->rev == 2)
526 max_width = 2048;
527
528 return max_width;
529}
530
531int tilcdc_crtc_mode_valid(struct drm_crtc *crtc, struct drm_display_mode *mode)
532{
533 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
534 unsigned int bandwidth;
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500535 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
Rob Clark16ea9752013-01-08 15:04:28 -0600536
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500537 /*
538 * check to see if the width is within the range that
539 * the LCD Controller physically supports
540 */
Rob Clark16ea9752013-01-08 15:04:28 -0600541 if (mode->hdisplay > tilcdc_crtc_max_width(crtc))
542 return MODE_VIRTUAL_X;
543
544 /* width must be multiple of 16 */
545 if (mode->hdisplay & 0xf)
546 return MODE_VIRTUAL_X;
547
548 if (mode->vdisplay > 2048)
549 return MODE_VIRTUAL_Y;
550
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500551 DBG("Processing mode %dx%d@%d with pixel clock %d",
552 mode->hdisplay, mode->vdisplay,
553 drm_mode_vrefresh(mode), mode->clock);
554
555 hbp = mode->htotal - mode->hsync_end;
556 hfp = mode->hsync_start - mode->hdisplay;
557 hsw = mode->hsync_end - mode->hsync_start;
558 vbp = mode->vtotal - mode->vsync_end;
559 vfp = mode->vsync_start - mode->vdisplay;
560 vsw = mode->vsync_end - mode->vsync_start;
561
562 if ((hbp-1) & ~0x3ff) {
563 DBG("Pruning mode: Horizontal Back Porch out of range");
564 return MODE_HBLANK_WIDE;
565 }
566
567 if ((hfp-1) & ~0x3ff) {
568 DBG("Pruning mode: Horizontal Front Porch out of range");
569 return MODE_HBLANK_WIDE;
570 }
571
572 if ((hsw-1) & ~0x3ff) {
573 DBG("Pruning mode: Horizontal Sync Width out of range");
574 return MODE_HSYNC_WIDE;
575 }
576
577 if (vbp & ~0xff) {
578 DBG("Pruning mode: Vertical Back Porch out of range");
579 return MODE_VBLANK_WIDE;
580 }
581
582 if (vfp & ~0xff) {
583 DBG("Pruning mode: Vertical Front Porch out of range");
584 return MODE_VBLANK_WIDE;
585 }
586
587 if ((vsw-1) & ~0x3f) {
588 DBG("Pruning mode: Vertical Sync Width out of range");
589 return MODE_VSYNC_WIDE;
590 }
591
Darren Etheridge4e564342013-06-21 13:52:23 -0500592 /*
593 * some devices have a maximum allowed pixel clock
594 * configured from the DT
595 */
596 if (mode->clock > priv->max_pixelclock) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500597 DBG("Pruning mode: pixel clock too high");
Darren Etheridge4e564342013-06-21 13:52:23 -0500598 return MODE_CLOCK_HIGH;
599 }
600
601 /*
602 * some devices further limit the max horizontal resolution
603 * configured from the DT
604 */
605 if (mode->hdisplay > priv->max_width)
606 return MODE_BAD_WIDTH;
607
Rob Clark16ea9752013-01-08 15:04:28 -0600608 /* filter out modes that would require too much memory bandwidth: */
Darren Etheridge4e564342013-06-21 13:52:23 -0500609 bandwidth = mode->hdisplay * mode->vdisplay *
610 drm_mode_vrefresh(mode);
611 if (bandwidth > priv->max_bandwidth) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500612 DBG("Pruning mode: exceeds defined bandwidth limit");
Rob Clark16ea9752013-01-08 15:04:28 -0600613 return MODE_BAD;
Darren Etheridge4e564342013-06-21 13:52:23 -0500614 }
Rob Clark16ea9752013-01-08 15:04:28 -0600615
616 return MODE_OK;
617}
618
619void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
620 const struct tilcdc_panel_info *info)
621{
622 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
623 tilcdc_crtc->info = info;
624}
625
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200626void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
627 bool simulate_vesa_sync)
628{
629 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
630
631 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
632}
633
Rob Clark16ea9752013-01-08 15:04:28 -0600634void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
635{
Rob Clark16ea9752013-01-08 15:04:28 -0600636 struct drm_device *dev = crtc->dev;
637 struct tilcdc_drm_private *priv = dev->dev_private;
Darren Etheridge3d193062014-01-15 15:52:36 -0600638 unsigned long lcd_clk;
639 const unsigned clkdiv = 2; /* using a fixed divider of 2 */
Rob Clark16ea9752013-01-08 15:04:28 -0600640 int ret;
641
642 pm_runtime_get_sync(dev->dev);
643
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300644 tilcdc_crtc_disable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600645
Darren Etheridge3d193062014-01-15 15:52:36 -0600646 /* mode.clock is in KHz, set_rate wants parameter in Hz */
647 ret = clk_set_rate(priv->clk, crtc->mode.clock * 1000 * clkdiv);
648 if (ret < 0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600649 dev_err(dev->dev, "failed to set display clock rate to: %d\n",
650 crtc->mode.clock);
651 goto out;
652 }
653
654 lcd_clk = clk_get_rate(priv->clk);
Rob Clark16ea9752013-01-08 15:04:28 -0600655
Darren Etheridge3d193062014-01-15 15:52:36 -0600656 DBG("lcd_clk=%lu, mode clock=%d, div=%u",
657 lcd_clk, crtc->mode.clock, clkdiv);
Rob Clark16ea9752013-01-08 15:04:28 -0600658
659 /* Configure the LCD clock divisor. */
Darren Etheridge3d193062014-01-15 15:52:36 -0600660 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
Rob Clark16ea9752013-01-08 15:04:28 -0600661 LCDC_RASTER_MODE);
662
663 if (priv->rev == 2)
664 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
665 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
666 LCDC_V2_CORE_CLK_EN);
667
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300668 if (tilcdc_crtc_is_on(crtc))
669 tilcdc_crtc_enable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600670
671out:
672 pm_runtime_put_sync(dev->dev);
673}
674
Jyri Sarha5895d082016-01-08 14:33:09 +0200675#define SYNC_LOST_COUNT_LIMIT 50
676
Rob Clark16ea9752013-01-08 15:04:28 -0600677irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
678{
679 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
680 struct drm_device *dev = crtc->dev;
681 struct tilcdc_drm_private *priv = dev->dev_private;
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300682 uint32_t stat;
Rob Clark16ea9752013-01-08 15:04:28 -0600683
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300684 stat = tilcdc_read_irqstatus(dev);
685 tilcdc_clear_irqstatus(dev, stat);
686
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300687 if (stat & LCDC_END_OF_FRAME0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600688 unsigned long flags;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200689 bool skip_event = false;
690 ktime_t now;
691
692 now = ktime_get();
Rob Clark16ea9752013-01-08 15:04:28 -0600693
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300694 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600695
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200696 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600697
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200698 tilcdc_crtc->last_vblank = now;
Rob Clark16ea9752013-01-08 15:04:28 -0600699
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200700 if (tilcdc_crtc->next_fb) {
701 set_scanout(crtc, tilcdc_crtc->next_fb);
702 tilcdc_crtc->next_fb = NULL;
703 skip_event = true;
Tomi Valkeinen2b2080d2015-10-20 09:37:27 +0300704 }
705
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200706 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
707
Gustavo Padovan099ede82016-07-04 21:04:52 -0300708 drm_crtc_handle_vblank(crtc);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200709
710 if (!skip_event) {
711 struct drm_pending_vblank_event *event;
712
713 spin_lock_irqsave(&dev->event_lock, flags);
714
715 event = tilcdc_crtc->event;
716 tilcdc_crtc->event = NULL;
717 if (event)
Gustavo Padovandfebc152016-04-14 10:48:22 -0700718 drm_crtc_send_vblank_event(crtc, event);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200719
720 spin_unlock_irqrestore(&dev->event_lock, flags);
721 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200722
723 if (tilcdc_crtc->frame_intact)
724 tilcdc_crtc->sync_lost_count = 0;
725 else
726 tilcdc_crtc->frame_intact = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600727 }
728
Jyri Sarha14944112016-04-07 20:36:48 +0300729 if (stat & LCDC_FIFO_UNDERFLOW)
730 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underfow",
731 __func__, stat);
732
733 /* For revision 2 only */
Rob Clark16ea9752013-01-08 15:04:28 -0600734 if (priv->rev == 2) {
735 if (stat & LCDC_FRAME_DONE) {
736 tilcdc_crtc->frame_done = true;
737 wake_up(&tilcdc_crtc->frame_done_wq);
738 }
Rob Clark16ea9752013-01-08 15:04:28 -0600739
Jyri Sarha1abcdac2016-06-17 11:54:06 +0300740 if (stat & LCDC_SYNC_LOST) {
741 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
742 __func__, stat);
743 tilcdc_crtc->frame_intact = false;
744 if (tilcdc_crtc->sync_lost_count++ >
745 SYNC_LOST_COUNT_LIMIT) {
746 dev_err(dev->dev, "%s(0x%08x): Sync lost flood detected, disabling the interrupt", __func__, stat);
747 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
748 LCDC_SYNC_LOST);
749 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200750 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200751
Jyri Sarha14944112016-04-07 20:36:48 +0300752 /* Indicate to LCDC that the interrupt service routine has
753 * completed, see 13.3.6.1.6 in AM335x TRM.
754 */
755 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
756 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200757
Rob Clark16ea9752013-01-08 15:04:28 -0600758 return IRQ_HANDLED;
759}
760
Rob Clark16ea9752013-01-08 15:04:28 -0600761struct drm_crtc *tilcdc_crtc_create(struct drm_device *dev)
762{
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300763 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600764 struct tilcdc_crtc *tilcdc_crtc;
765 struct drm_crtc *crtc;
766 int ret;
767
Jyri Sarhad0ec32c2016-02-23 12:44:27 +0200768 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
Rob Clark16ea9752013-01-08 15:04:28 -0600769 if (!tilcdc_crtc) {
770 dev_err(dev->dev, "allocation failed\n");
771 return NULL;
772 }
773
774 crtc = &tilcdc_crtc->base;
775
Jyri Sarha47f571c2016-04-07 15:04:18 +0300776 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
777 if (ret < 0)
778 goto fail;
779
Rob Clark16ea9752013-01-08 15:04:28 -0600780 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
781
Boris BREZILLONd7f8db52014-11-14 19:30:30 +0100782 drm_flip_work_init(&tilcdc_crtc->unref_work,
Rob Clarka464d612013-08-07 13:41:20 -0400783 "unref", unref_worker);
Rob Clark16ea9752013-01-08 15:04:28 -0600784
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200785 spin_lock_init(&tilcdc_crtc->irq_lock);
786
Jyri Sarha47f571c2016-04-07 15:04:18 +0300787 ret = drm_crtc_init_with_planes(dev, crtc,
788 &tilcdc_crtc->primary,
789 NULL,
790 &tilcdc_crtc_funcs,
791 "tilcdc crtc");
Rob Clark16ea9752013-01-08 15:04:28 -0600792 if (ret < 0)
793 goto fail;
794
795 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
796
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300797 if (priv->is_componentized) {
798 struct device_node *ports =
799 of_get_child_by_name(dev->dev->of_node, "ports");
800
801 if (ports) {
802 crtc->port = of_get_child_by_name(ports, "port");
803 of_node_put(ports);
804 } else {
805 crtc->port =
806 of_get_child_by_name(dev->dev->of_node, "port");
807 }
808 if (!crtc->port) { /* This should never happen */
809 dev_err(dev->dev, "Port node not found in %s\n",
810 dev->dev->of_node->full_name);
811 goto fail;
812 }
813 }
814
Rob Clark16ea9752013-01-08 15:04:28 -0600815 return crtc;
816
817fail:
818 tilcdc_crtc_destroy(crtc);
819 return NULL;
820}