blob: 773bee27a2d08aaee09af919a2b7c9275a6b0cda [file] [log] [blame]
Rob Clark16ea9752013-01-08 15:04:28 -06001/*
2 * Copyright (C) 2012 Texas Instruments
3 * Author: Rob Clark <robdclark@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
16 */
17
Rob Clarka464d612013-08-07 13:41:20 -040018#include "drm_flip_work.h"
Daniel Vetter3cb9ae42014-10-29 10:03:57 +010019#include <drm/drm_plane_helper.h>
Jyri Sarha305198d2016-04-07 15:05:16 +030020#include <drm/drm_atomic_helper.h>
Rob Clark16ea9752013-01-08 15:04:28 -060021
22#include "tilcdc_drv.h"
23#include "tilcdc_regs.h"
24
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020025#define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
26
Rob Clark16ea9752013-01-08 15:04:28 -060027struct tilcdc_crtc {
28 struct drm_crtc base;
29
Jyri Sarha47f571c2016-04-07 15:04:18 +030030 struct drm_plane primary;
Rob Clark16ea9752013-01-08 15:04:28 -060031 const struct tilcdc_panel_info *info;
Rob Clark16ea9752013-01-08 15:04:28 -060032 struct drm_pending_vblank_event *event;
33 int dpms;
34 wait_queue_head_t frame_done_wq;
35 bool frame_done;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020036 spinlock_t irq_lock;
37
38 ktime_t last_vblank;
Rob Clark16ea9752013-01-08 15:04:28 -060039
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030040 struct drm_framebuffer *curr_fb;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020041 struct drm_framebuffer *next_fb;
Rob Clark16ea9752013-01-08 15:04:28 -060042
43 /* for deferred fb unref's: */
Rob Clarka464d612013-08-07 13:41:20 -040044 struct drm_flip_work unref_work;
Jyri Sarha103cd8b2015-02-10 14:13:23 +020045
46 /* Only set if an external encoder is connected */
47 bool simulate_vesa_sync;
Jyri Sarha5895d082016-01-08 14:33:09 +020048
49 int sync_lost_count;
50 bool frame_intact;
Rob Clark16ea9752013-01-08 15:04:28 -060051};
52#define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
53
Rob Clarka464d612013-08-07 13:41:20 -040054static void unref_worker(struct drm_flip_work *work, void *val)
Rob Clark16ea9752013-01-08 15:04:28 -060055{
Darren Etheridgef7b45752013-06-21 13:52:26 -050056 struct tilcdc_crtc *tilcdc_crtc =
Rob Clarka464d612013-08-07 13:41:20 -040057 container_of(work, struct tilcdc_crtc, unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -060058 struct drm_device *dev = tilcdc_crtc->base.dev;
Rob Clark16ea9752013-01-08 15:04:28 -060059
60 mutex_lock(&dev->mode_config.mutex);
Rob Clarka464d612013-08-07 13:41:20 -040061 drm_framebuffer_unreference(val);
Rob Clark16ea9752013-01-08 15:04:28 -060062 mutex_unlock(&dev->mode_config.mutex);
63}
64
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030065static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
Rob Clark16ea9752013-01-08 15:04:28 -060066{
67 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
68 struct drm_device *dev = crtc->dev;
Rob Clark16ea9752013-01-08 15:04:28 -060069 struct drm_gem_cma_object *gem;
70 unsigned int depth, bpp;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030071 dma_addr_t start, end;
Rob Clark16ea9752013-01-08 15:04:28 -060072
73 drm_fb_get_bpp_depth(fb->pixel_format, &depth, &bpp);
74 gem = drm_fb_cma_get_gem_obj(fb, 0);
75
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030076 start = gem->paddr + fb->offsets[0] +
77 crtc->y * fb->pitches[0] +
78 crtc->x * bpp / 8;
Rob Clark16ea9752013-01-08 15:04:28 -060079
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030080 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
Rob Clark16ea9752013-01-08 15:04:28 -060081
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030082 tilcdc_write(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, start);
83 tilcdc_write(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG, end);
84
85 if (tilcdc_crtc->curr_fb)
86 drm_flip_work_queue(&tilcdc_crtc->unref_work,
87 tilcdc_crtc->curr_fb);
88
89 tilcdc_crtc->curr_fb = fb;
Rob Clark16ea9752013-01-08 15:04:28 -060090}
91
Jyri Sarhaafaf8332016-06-21 16:00:44 +030092static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
93{
94 struct tilcdc_drm_private *priv = dev->dev_private;
95
96 tilcdc_clear_irqstatus(dev, 0xffffffff);
97
98 if (priv->rev == 1) {
99 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
100 LCDC_V1_UNDERFLOW_INT_ENA);
101 } else {
102 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
103 LCDC_V2_UNDERFLOW_INT_ENA |
104 LCDC_V2_END_OF_FRAME0_INT_ENA |
105 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
106 }
107}
108
109static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
110{
111 struct tilcdc_drm_private *priv = dev->dev_private;
112
113 /* disable irqs that we might have enabled: */
114 if (priv->rev == 1) {
115 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
116 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
117 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
118 LCDC_V1_END_OF_FRAME_INT_ENA);
119 } else {
120 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
121 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
122 LCDC_V2_END_OF_FRAME0_INT_ENA |
123 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
124 }
125}
126
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300127static void reset(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600128{
129 struct drm_device *dev = crtc->dev;
130 struct tilcdc_drm_private *priv = dev->dev_private;
131
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300132 if (priv->rev != 2)
133 return;
134
135 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
136 usleep_range(250, 1000);
137 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
138}
139
140static void start(struct drm_crtc *crtc)
141{
142 struct drm_device *dev = crtc->dev;
143
144 reset(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600145
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300146 tilcdc_crtc_enable_irqs(dev);
147
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300148 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
Rob Clark16ea9752013-01-08 15:04:28 -0600149 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_PALETTE_LOAD_MODE(DATA_ONLY));
150 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300151
152 drm_crtc_vblank_on(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600153}
154
155static void stop(struct drm_crtc *crtc)
156{
Jyri Sarha2d5be882016-04-07 20:20:23 +0300157 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600158 struct drm_device *dev = crtc->dev;
Jyri Sarha2d5be882016-04-07 20:20:23 +0300159 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600160
Jyri Sarha2d5be882016-04-07 20:20:23 +0300161 tilcdc_crtc->frame_done = false;
Rob Clark16ea9752013-01-08 15:04:28 -0600162 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha2d5be882016-04-07 20:20:23 +0300163
164 /*
165 * if necessary wait for framedone irq which will still come
166 * before putting things to sleep..
167 */
168 if (priv->rev == 2) {
169 int ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
170 tilcdc_crtc->frame_done,
Jyri Sarha437c7d92016-06-16 16:19:17 +0300171 msecs_to_jiffies(500));
Jyri Sarha2d5be882016-04-07 20:20:23 +0300172 if (ret == 0)
173 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
174 __func__);
175 }
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300176
177 drm_crtc_vblank_off(crtc);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300178
179 tilcdc_crtc_disable_irqs(dev);
Rob Clark16ea9752013-01-08 15:04:28 -0600180}
181
182static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
183{
184 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
185
Jyri Sarhade9cb5f2015-02-26 10:12:41 +0200186 tilcdc_crtc_dpms(crtc, DRM_MODE_DPMS_OFF);
Rob Clark16ea9752013-01-08 15:04:28 -0600187
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300188 of_node_put(crtc->port);
Rob Clark16ea9752013-01-08 15:04:28 -0600189 drm_crtc_cleanup(crtc);
Rob Clarka464d612013-08-07 13:41:20 -0400190 drm_flip_work_cleanup(&tilcdc_crtc->unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -0600191}
192
Jyri Sarha8c65abb2016-04-07 14:56:32 +0300193int tilcdc_crtc_page_flip(struct drm_crtc *crtc,
Rob Clark16ea9752013-01-08 15:04:28 -0600194 struct drm_framebuffer *fb,
Keith Packarded8d1972013-07-22 18:49:58 -0700195 struct drm_pending_vblank_event *event,
196 uint32_t page_flip_flags)
Rob Clark16ea9752013-01-08 15:04:28 -0600197{
198 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
199 struct drm_device *dev = crtc->dev;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300200 unsigned long flags;
Tomi Valkeinen6f206e92014-02-07 17:37:07 +0000201
Rob Clark16ea9752013-01-08 15:04:28 -0600202 if (tilcdc_crtc->event) {
203 dev_err(dev->dev, "already pending page flip!\n");
204 return -EBUSY;
205 }
206
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300207 drm_framebuffer_reference(fb);
208
Matt Roperf4510a22014-04-01 15:22:40 -0700209 crtc->primary->fb = fb;
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300210
211 pm_runtime_get_sync(dev->dev);
212
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200213 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300214
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300215 if (crtc->hwmode.vrefresh && ktime_to_ns(tilcdc_crtc->last_vblank)) {
216 ktime_t next_vblank;
217 s64 tdiff;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300218
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300219 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
220 1000000 / crtc->hwmode.vrefresh);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200221
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300222 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
223
224 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
225 tilcdc_crtc->next_fb = fb;
226 }
227
228 if (tilcdc_crtc->next_fb != fb)
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200229 set_scanout(crtc, fb);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200230
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300231 tilcdc_crtc->event = event;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200232
233 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600234
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300235 pm_runtime_put_sync(dev->dev);
236
Rob Clark16ea9752013-01-08 15:04:28 -0600237 return 0;
238}
239
Darren Etheridge614b3cfe2014-09-25 00:59:32 +0000240void tilcdc_crtc_dpms(struct drm_crtc *crtc, int mode)
Rob Clark16ea9752013-01-08 15:04:28 -0600241{
242 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
243 struct drm_device *dev = crtc->dev;
244 struct tilcdc_drm_private *priv = dev->dev_private;
245
246 /* we really only care about on or off: */
247 if (mode != DRM_MODE_DPMS_ON)
248 mode = DRM_MODE_DPMS_OFF;
249
250 if (tilcdc_crtc->dpms == mode)
251 return;
252
253 tilcdc_crtc->dpms = mode;
254
Rob Clark16ea9752013-01-08 15:04:28 -0600255 if (mode == DRM_MODE_DPMS_ON) {
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300256 pm_runtime_get_sync(dev->dev);
Rob Clark16ea9752013-01-08 15:04:28 -0600257 start(crtc);
258 } else {
Rob Clark16ea9752013-01-08 15:04:28 -0600259 stop(crtc);
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300260 pm_runtime_put_sync(dev->dev);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300261
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200262 if (tilcdc_crtc->next_fb) {
263 drm_flip_work_queue(&tilcdc_crtc->unref_work,
264 tilcdc_crtc->next_fb);
265 tilcdc_crtc->next_fb = NULL;
266 }
267
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300268 if (tilcdc_crtc->curr_fb) {
269 drm_flip_work_queue(&tilcdc_crtc->unref_work,
270 tilcdc_crtc->curr_fb);
271 tilcdc_crtc->curr_fb = NULL;
272 }
273
274 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300275 tilcdc_crtc->last_vblank = ktime_set(0, 0);
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300276 }
Rob Clark16ea9752013-01-08 15:04:28 -0600277}
278
Jyri Sarha8fe56162016-06-14 11:43:30 +0300279int tilcdc_crtc_current_dpms_state(struct drm_crtc *crtc)
280{
281 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
282
283 return tilcdc_crtc->dpms;
284}
285
Rob Clark16ea9752013-01-08 15:04:28 -0600286static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
287 const struct drm_display_mode *mode,
288 struct drm_display_mode *adjusted_mode)
289{
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200290 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
291
292 if (!tilcdc_crtc->simulate_vesa_sync)
293 return true;
294
295 /*
296 * tilcdc does not generate VESA-compliant sync but aligns
297 * VS on the second edge of HS instead of first edge.
298 * We use adjusted_mode, to fixup sync by aligning both rising
299 * edges and add HSKEW offset to fix the sync.
300 */
301 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
302 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
303
304 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
305 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
306 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
307 } else {
308 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
309 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
310 }
311
Rob Clark16ea9752013-01-08 15:04:28 -0600312 return true;
313}
314
Jyri Sarha305198d2016-04-07 15:05:16 +0300315static void tilcdc_crtc_disable(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600316{
317 tilcdc_crtc_dpms(crtc, DRM_MODE_DPMS_OFF);
318}
319
Jyri Sarha305198d2016-04-07 15:05:16 +0300320static void tilcdc_crtc_enable(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600321{
322 tilcdc_crtc_dpms(crtc, DRM_MODE_DPMS_ON);
323}
324
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300325static void tilcdc_crtc_mode_set_nofb(struct drm_crtc *crtc)
326{
327 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
328 struct drm_device *dev = crtc->dev;
329 struct tilcdc_drm_private *priv = dev->dev_private;
330 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
331 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
332 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
333 struct drm_framebuffer *fb = crtc->primary->state->fb;
334
335 if (WARN_ON(!info))
336 return;
337
338 if (WARN_ON(!fb))
339 return;
340
341 pm_runtime_get_sync(dev->dev);
342
343 /* Configure the Burst Size and fifo threshold of DMA: */
344 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
345 switch (info->dma_burst_sz) {
346 case 1:
347 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
348 break;
349 case 2:
350 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
351 break;
352 case 4:
353 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
354 break;
355 case 8:
356 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
357 break;
358 case 16:
359 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
360 break;
361 default:
362 dev_err(dev->dev, "invalid burst size\n");
363 return;
364 }
365 reg |= (info->fifo_th << 8);
366 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
367
368 /* Configure timings: */
369 hbp = mode->htotal - mode->hsync_end;
370 hfp = mode->hsync_start - mode->hdisplay;
371 hsw = mode->hsync_end - mode->hsync_start;
372 vbp = mode->vtotal - mode->vsync_end;
373 vfp = mode->vsync_start - mode->vdisplay;
374 vsw = mode->vsync_end - mode->vsync_start;
375
376 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
377 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
378
379 /* Set AC Bias Period and Number of Transitions per Interrupt: */
380 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
381 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
382 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
383
384 /*
385 * subtract one from hfp, hbp, hsw because the hardware uses
386 * a value of 0 as 1
387 */
388 if (priv->rev == 2) {
389 /* clear bits we're going to set */
390 reg &= ~0x78000033;
391 reg |= ((hfp-1) & 0x300) >> 8;
392 reg |= ((hbp-1) & 0x300) >> 4;
393 reg |= ((hsw-1) & 0x3c0) << 21;
394 }
395 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
396
397 reg = (((mode->hdisplay >> 4) - 1) << 4) |
398 (((hbp-1) & 0xff) << 24) |
399 (((hfp-1) & 0xff) << 16) |
400 (((hsw-1) & 0x3f) << 10);
401 if (priv->rev == 2)
402 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
403 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
404
405 reg = ((mode->vdisplay - 1) & 0x3ff) |
406 ((vbp & 0xff) << 24) |
407 ((vfp & 0xff) << 16) |
408 (((vsw-1) & 0x3f) << 10);
409 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
410
411 /*
412 * be sure to set Bit 10 for the V2 LCDC controller,
413 * otherwise limited to 1024 pixels width, stopping
414 * 1920x1080 being supported.
415 */
416 if (priv->rev == 2) {
417 if ((mode->vdisplay - 1) & 0x400) {
418 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
419 LCDC_LPP_B10);
420 } else {
421 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
422 LCDC_LPP_B10);
423 }
424 }
425
426 /* Configure display type: */
427 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
428 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
429 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
430 0x000ff000 /* Palette Loading Delay bits */);
431 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
432 if (info->tft_alt_mode)
433 reg |= LCDC_TFT_ALT_ENABLE;
434 if (priv->rev == 2) {
435 unsigned int depth, bpp;
436
437 drm_fb_get_bpp_depth(fb->pixel_format, &depth, &bpp);
438 switch (bpp) {
439 case 16:
440 break;
441 case 32:
442 reg |= LCDC_V2_TFT_24BPP_UNPACK;
443 /* fallthrough */
444 case 24:
445 reg |= LCDC_V2_TFT_24BPP_MODE;
446 break;
447 default:
448 dev_err(dev->dev, "invalid pixel format\n");
449 return;
450 }
451 }
452 reg |= info->fdd < 12;
453 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
454
455 if (info->invert_pxl_clk)
456 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
457 else
458 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
459
460 if (info->sync_ctrl)
461 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
462 else
463 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
464
465 if (info->sync_edge)
466 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
467 else
468 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
469
470 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
471 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
472 else
473 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
474
475 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
476 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
477 else
478 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
479
480 if (info->raster_order)
481 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
482 else
483 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
484
485 drm_framebuffer_reference(fb);
486
487 set_scanout(crtc, fb);
488
489 tilcdc_crtc_update_clk(crtc);
490
491 pm_runtime_put_sync(dev->dev);
492
493 crtc->hwmode = crtc->state->adjusted_mode;
494}
495
Jyri Sarhadb380c52016-04-07 15:10:23 +0300496static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
497 struct drm_crtc_state *state)
498{
499 struct drm_display_mode *mode = &state->mode;
500 int ret;
501
502 /* If we are not active we don't care */
503 if (!state->active)
504 return 0;
505
506 if (state->state->planes[0].ptr != crtc->primary ||
507 state->state->planes[0].state == NULL ||
508 state->state->planes[0].state->crtc != crtc) {
509 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
510 return -EINVAL;
511 }
512
513 ret = tilcdc_crtc_mode_valid(crtc, mode);
514 if (ret) {
515 dev_dbg(crtc->dev->dev, "Mode \"%s\" not valid", mode->name);
516 return -EINVAL;
517 }
518
519 return 0;
520}
521
Rob Clark16ea9752013-01-08 15:04:28 -0600522static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
Jyri Sarha305198d2016-04-07 15:05:16 +0300523 .destroy = tilcdc_crtc_destroy,
524 .set_config = drm_atomic_helper_set_config,
525 .page_flip = drm_atomic_helper_page_flip,
526 .reset = drm_atomic_helper_crtc_reset,
527 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
528 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
Rob Clark16ea9752013-01-08 15:04:28 -0600529};
530
531static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
Rob Clark16ea9752013-01-08 15:04:28 -0600532 .mode_fixup = tilcdc_crtc_mode_fixup,
Jyri Sarha305198d2016-04-07 15:05:16 +0300533 .enable = tilcdc_crtc_enable,
534 .disable = tilcdc_crtc_disable,
Jyri Sarhadb380c52016-04-07 15:10:23 +0300535 .atomic_check = tilcdc_crtc_atomic_check,
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300536 .mode_set_nofb = tilcdc_crtc_mode_set_nofb,
Rob Clark16ea9752013-01-08 15:04:28 -0600537};
538
539int tilcdc_crtc_max_width(struct drm_crtc *crtc)
540{
541 struct drm_device *dev = crtc->dev;
542 struct tilcdc_drm_private *priv = dev->dev_private;
543 int max_width = 0;
544
545 if (priv->rev == 1)
546 max_width = 1024;
547 else if (priv->rev == 2)
548 max_width = 2048;
549
550 return max_width;
551}
552
553int tilcdc_crtc_mode_valid(struct drm_crtc *crtc, struct drm_display_mode *mode)
554{
555 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
556 unsigned int bandwidth;
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500557 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
Rob Clark16ea9752013-01-08 15:04:28 -0600558
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500559 /*
560 * check to see if the width is within the range that
561 * the LCD Controller physically supports
562 */
Rob Clark16ea9752013-01-08 15:04:28 -0600563 if (mode->hdisplay > tilcdc_crtc_max_width(crtc))
564 return MODE_VIRTUAL_X;
565
566 /* width must be multiple of 16 */
567 if (mode->hdisplay & 0xf)
568 return MODE_VIRTUAL_X;
569
570 if (mode->vdisplay > 2048)
571 return MODE_VIRTUAL_Y;
572
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500573 DBG("Processing mode %dx%d@%d with pixel clock %d",
574 mode->hdisplay, mode->vdisplay,
575 drm_mode_vrefresh(mode), mode->clock);
576
577 hbp = mode->htotal - mode->hsync_end;
578 hfp = mode->hsync_start - mode->hdisplay;
579 hsw = mode->hsync_end - mode->hsync_start;
580 vbp = mode->vtotal - mode->vsync_end;
581 vfp = mode->vsync_start - mode->vdisplay;
582 vsw = mode->vsync_end - mode->vsync_start;
583
584 if ((hbp-1) & ~0x3ff) {
585 DBG("Pruning mode: Horizontal Back Porch out of range");
586 return MODE_HBLANK_WIDE;
587 }
588
589 if ((hfp-1) & ~0x3ff) {
590 DBG("Pruning mode: Horizontal Front Porch out of range");
591 return MODE_HBLANK_WIDE;
592 }
593
594 if ((hsw-1) & ~0x3ff) {
595 DBG("Pruning mode: Horizontal Sync Width out of range");
596 return MODE_HSYNC_WIDE;
597 }
598
599 if (vbp & ~0xff) {
600 DBG("Pruning mode: Vertical Back Porch out of range");
601 return MODE_VBLANK_WIDE;
602 }
603
604 if (vfp & ~0xff) {
605 DBG("Pruning mode: Vertical Front Porch out of range");
606 return MODE_VBLANK_WIDE;
607 }
608
609 if ((vsw-1) & ~0x3f) {
610 DBG("Pruning mode: Vertical Sync Width out of range");
611 return MODE_VSYNC_WIDE;
612 }
613
Darren Etheridge4e564342013-06-21 13:52:23 -0500614 /*
615 * some devices have a maximum allowed pixel clock
616 * configured from the DT
617 */
618 if (mode->clock > priv->max_pixelclock) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500619 DBG("Pruning mode: pixel clock too high");
Darren Etheridge4e564342013-06-21 13:52:23 -0500620 return MODE_CLOCK_HIGH;
621 }
622
623 /*
624 * some devices further limit the max horizontal resolution
625 * configured from the DT
626 */
627 if (mode->hdisplay > priv->max_width)
628 return MODE_BAD_WIDTH;
629
Rob Clark16ea9752013-01-08 15:04:28 -0600630 /* filter out modes that would require too much memory bandwidth: */
Darren Etheridge4e564342013-06-21 13:52:23 -0500631 bandwidth = mode->hdisplay * mode->vdisplay *
632 drm_mode_vrefresh(mode);
633 if (bandwidth > priv->max_bandwidth) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500634 DBG("Pruning mode: exceeds defined bandwidth limit");
Rob Clark16ea9752013-01-08 15:04:28 -0600635 return MODE_BAD;
Darren Etheridge4e564342013-06-21 13:52:23 -0500636 }
Rob Clark16ea9752013-01-08 15:04:28 -0600637
638 return MODE_OK;
639}
640
641void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
642 const struct tilcdc_panel_info *info)
643{
644 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
645 tilcdc_crtc->info = info;
646}
647
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200648void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
649 bool simulate_vesa_sync)
650{
651 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
652
653 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
654}
655
Rob Clark16ea9752013-01-08 15:04:28 -0600656void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
657{
658 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
659 struct drm_device *dev = crtc->dev;
660 struct tilcdc_drm_private *priv = dev->dev_private;
661 int dpms = tilcdc_crtc->dpms;
Darren Etheridge3d193062014-01-15 15:52:36 -0600662 unsigned long lcd_clk;
663 const unsigned clkdiv = 2; /* using a fixed divider of 2 */
Rob Clark16ea9752013-01-08 15:04:28 -0600664 int ret;
665
666 pm_runtime_get_sync(dev->dev);
667
668 if (dpms == DRM_MODE_DPMS_ON)
669 tilcdc_crtc_dpms(crtc, DRM_MODE_DPMS_OFF);
670
Darren Etheridge3d193062014-01-15 15:52:36 -0600671 /* mode.clock is in KHz, set_rate wants parameter in Hz */
672 ret = clk_set_rate(priv->clk, crtc->mode.clock * 1000 * clkdiv);
673 if (ret < 0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600674 dev_err(dev->dev, "failed to set display clock rate to: %d\n",
675 crtc->mode.clock);
676 goto out;
677 }
678
679 lcd_clk = clk_get_rate(priv->clk);
Rob Clark16ea9752013-01-08 15:04:28 -0600680
Darren Etheridge3d193062014-01-15 15:52:36 -0600681 DBG("lcd_clk=%lu, mode clock=%d, div=%u",
682 lcd_clk, crtc->mode.clock, clkdiv);
Rob Clark16ea9752013-01-08 15:04:28 -0600683
684 /* Configure the LCD clock divisor. */
Darren Etheridge3d193062014-01-15 15:52:36 -0600685 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
Rob Clark16ea9752013-01-08 15:04:28 -0600686 LCDC_RASTER_MODE);
687
688 if (priv->rev == 2)
689 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
690 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
691 LCDC_V2_CORE_CLK_EN);
692
693 if (dpms == DRM_MODE_DPMS_ON)
694 tilcdc_crtc_dpms(crtc, DRM_MODE_DPMS_ON);
695
696out:
697 pm_runtime_put_sync(dev->dev);
698}
699
Jyri Sarha5895d082016-01-08 14:33:09 +0200700#define SYNC_LOST_COUNT_LIMIT 50
701
Rob Clark16ea9752013-01-08 15:04:28 -0600702irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
703{
704 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
705 struct drm_device *dev = crtc->dev;
706 struct tilcdc_drm_private *priv = dev->dev_private;
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300707 uint32_t stat;
Rob Clark16ea9752013-01-08 15:04:28 -0600708
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300709 stat = tilcdc_read_irqstatus(dev);
710 tilcdc_clear_irqstatus(dev, stat);
711
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300712 if (stat & LCDC_END_OF_FRAME0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600713 unsigned long flags;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200714 bool skip_event = false;
715 ktime_t now;
716
717 now = ktime_get();
Rob Clark16ea9752013-01-08 15:04:28 -0600718
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300719 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600720
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200721 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600722
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200723 tilcdc_crtc->last_vblank = now;
Rob Clark16ea9752013-01-08 15:04:28 -0600724
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200725 if (tilcdc_crtc->next_fb) {
726 set_scanout(crtc, tilcdc_crtc->next_fb);
727 tilcdc_crtc->next_fb = NULL;
728 skip_event = true;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300729 }
730
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200731 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
732
Gustavo Padovan099ede82016-07-04 21:04:52 -0300733 drm_crtc_handle_vblank(crtc);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200734
735 if (!skip_event) {
736 struct drm_pending_vblank_event *event;
737
738 spin_lock_irqsave(&dev->event_lock, flags);
739
740 event = tilcdc_crtc->event;
741 tilcdc_crtc->event = NULL;
742 if (event)
Gustavo Padovandfebc152016-04-14 10:48:22 -0700743 drm_crtc_send_vblank_event(crtc, event);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200744
745 spin_unlock_irqrestore(&dev->event_lock, flags);
746 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200747
748 if (tilcdc_crtc->frame_intact)
749 tilcdc_crtc->sync_lost_count = 0;
750 else
751 tilcdc_crtc->frame_intact = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600752 }
753
Jyri Sarha14944112016-04-07 20:36:48 +0300754 if (stat & LCDC_FIFO_UNDERFLOW)
755 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underfow",
756 __func__, stat);
757
758 /* For revision 2 only */
Rob Clark16ea9752013-01-08 15:04:28 -0600759 if (priv->rev == 2) {
760 if (stat & LCDC_FRAME_DONE) {
761 tilcdc_crtc->frame_done = true;
762 wake_up(&tilcdc_crtc->frame_done_wq);
763 }
Rob Clark16ea9752013-01-08 15:04:28 -0600764
Jyri Sarha1abcdac2016-06-17 11:54:06 +0300765 if (stat & LCDC_SYNC_LOST) {
766 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
767 __func__, stat);
768 tilcdc_crtc->frame_intact = false;
769 if (tilcdc_crtc->sync_lost_count++ >
770 SYNC_LOST_COUNT_LIMIT) {
771 dev_err(dev->dev, "%s(0x%08x): Sync lost flood detected, disabling the interrupt", __func__, stat);
772 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
773 LCDC_SYNC_LOST);
774 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200775 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200776
Jyri Sarha14944112016-04-07 20:36:48 +0300777 /* Indicate to LCDC that the interrupt service routine has
778 * completed, see 13.3.6.1.6 in AM335x TRM.
779 */
780 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
781 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200782
Rob Clark16ea9752013-01-08 15:04:28 -0600783 return IRQ_HANDLED;
784}
785
Rob Clark16ea9752013-01-08 15:04:28 -0600786struct drm_crtc *tilcdc_crtc_create(struct drm_device *dev)
787{
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300788 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600789 struct tilcdc_crtc *tilcdc_crtc;
790 struct drm_crtc *crtc;
791 int ret;
792
Jyri Sarhad0ec32c2016-02-23 12:44:27 +0200793 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
Rob Clark16ea9752013-01-08 15:04:28 -0600794 if (!tilcdc_crtc) {
795 dev_err(dev->dev, "allocation failed\n");
796 return NULL;
797 }
798
799 crtc = &tilcdc_crtc->base;
800
Jyri Sarha47f571c2016-04-07 15:04:18 +0300801 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
802 if (ret < 0)
803 goto fail;
804
Rob Clark16ea9752013-01-08 15:04:28 -0600805 tilcdc_crtc->dpms = DRM_MODE_DPMS_OFF;
806 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
807
Boris BREZILLONd7f8db52014-11-14 19:30:30 +0100808 drm_flip_work_init(&tilcdc_crtc->unref_work,
Rob Clarka464d612013-08-07 13:41:20 -0400809 "unref", unref_worker);
Rob Clark16ea9752013-01-08 15:04:28 -0600810
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200811 spin_lock_init(&tilcdc_crtc->irq_lock);
812
Jyri Sarha47f571c2016-04-07 15:04:18 +0300813 ret = drm_crtc_init_with_planes(dev, crtc,
814 &tilcdc_crtc->primary,
815 NULL,
816 &tilcdc_crtc_funcs,
817 "tilcdc crtc");
Rob Clark16ea9752013-01-08 15:04:28 -0600818 if (ret < 0)
819 goto fail;
820
821 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
822
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300823 if (priv->is_componentized) {
824 struct device_node *ports =
825 of_get_child_by_name(dev->dev->of_node, "ports");
826
827 if (ports) {
828 crtc->port = of_get_child_by_name(ports, "port");
829 of_node_put(ports);
830 } else {
831 crtc->port =
832 of_get_child_by_name(dev->dev->of_node, "port");
833 }
834 if (!crtc->port) { /* This should never happen */
835 dev_err(dev->dev, "Port node not found in %s\n",
836 dev->dev->of_node->full_name);
837 goto fail;
838 }
839 }
840
Rob Clark16ea9752013-01-08 15:04:28 -0600841 return crtc;
842
843fail:
844 tilcdc_crtc_destroy(crtc);
845 return NULL;
846}