blob: 7ea34c21c3840076f3503ad9f8168bddf417cfdc [file] [log] [blame]
Rob Clark16ea9752013-01-08 15:04:28 -06001/*
2 * Copyright (C) 2012 Texas Instruments
3 * Author: Rob Clark <robdclark@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
16 */
17
Sean Paulce2f2c32016-09-21 06:14:53 -070018#include <drm/drm_atomic.h>
Jyri Sarha305198d2016-04-07 15:05:16 +030019#include <drm/drm_atomic_helper.h>
Sean Paulce2f2c32016-09-21 06:14:53 -070020#include <drm/drm_crtc.h>
21#include <drm/drm_flip_work.h>
22#include <drm/drm_plane_helper.h>
Jyri Sarha4e910c72016-09-06 22:55:33 +030023#include <linux/workqueue.h>
Bartosz Golaszewski93452352016-10-31 15:19:26 +010024#include <linux/completion.h>
25#include <linux/dma-mapping.h>
Rob Clark16ea9752013-01-08 15:04:28 -060026
27#include "tilcdc_drv.h"
28#include "tilcdc_regs.h"
29
Bartosz Golaszewski93452352016-10-31 15:19:26 +010030#define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
31#define TILCDC_REV1_PALETTE_SIZE 32
32#define TILCDC_REV1_PALETTE_FIRST_ENTRY 0x4000
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020033
Rob Clark16ea9752013-01-08 15:04:28 -060034struct tilcdc_crtc {
35 struct drm_crtc base;
36
Jyri Sarha47f571c2016-04-07 15:04:18 +030037 struct drm_plane primary;
Rob Clark16ea9752013-01-08 15:04:28 -060038 const struct tilcdc_panel_info *info;
Rob Clark16ea9752013-01-08 15:04:28 -060039 struct drm_pending_vblank_event *event;
Jyri Sarha2d53a182016-10-25 12:27:31 +030040 struct mutex enable_lock;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +030041 bool enabled;
Jyri Sarha2d53a182016-10-25 12:27:31 +030042 bool shutdown;
Rob Clark16ea9752013-01-08 15:04:28 -060043 wait_queue_head_t frame_done_wq;
44 bool frame_done;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020045 spinlock_t irq_lock;
46
Jyri Sarha642e5162016-09-06 16:19:54 +030047 unsigned int lcd_fck_rate;
48
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020049 ktime_t last_vblank;
Rob Clark16ea9752013-01-08 15:04:28 -060050
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030051 struct drm_framebuffer *curr_fb;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020052 struct drm_framebuffer *next_fb;
Rob Clark16ea9752013-01-08 15:04:28 -060053
54 /* for deferred fb unref's: */
Rob Clarka464d612013-08-07 13:41:20 -040055 struct drm_flip_work unref_work;
Jyri Sarha103cd8b2015-02-10 14:13:23 +020056
57 /* Only set if an external encoder is connected */
58 bool simulate_vesa_sync;
Jyri Sarha5895d082016-01-08 14:33:09 +020059
60 int sync_lost_count;
61 bool frame_intact;
Jyri Sarha13b3d722016-04-06 14:02:38 +030062 struct work_struct recover_work;
Bartosz Golaszewski93452352016-10-31 15:19:26 +010063
64 dma_addr_t palette_dma_handle;
65 void *palette_base;
66 struct completion palette_loaded;
Rob Clark16ea9752013-01-08 15:04:28 -060067};
68#define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
69
Rob Clarka464d612013-08-07 13:41:20 -040070static void unref_worker(struct drm_flip_work *work, void *val)
Rob Clark16ea9752013-01-08 15:04:28 -060071{
Darren Etheridgef7b45752013-06-21 13:52:26 -050072 struct tilcdc_crtc *tilcdc_crtc =
Rob Clarka464d612013-08-07 13:41:20 -040073 container_of(work, struct tilcdc_crtc, unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -060074 struct drm_device *dev = tilcdc_crtc->base.dev;
Rob Clark16ea9752013-01-08 15:04:28 -060075
76 mutex_lock(&dev->mode_config.mutex);
Rob Clarka464d612013-08-07 13:41:20 -040077 drm_framebuffer_unreference(val);
Rob Clark16ea9752013-01-08 15:04:28 -060078 mutex_unlock(&dev->mode_config.mutex);
79}
80
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030081static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
Rob Clark16ea9752013-01-08 15:04:28 -060082{
83 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
84 struct drm_device *dev = crtc->dev;
Daniel Schultz4c268d62016-10-28 13:52:41 +020085 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -060086 struct drm_gem_cma_object *gem;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030087 dma_addr_t start, end;
Jyri Sarha7eb9f062016-08-26 15:10:14 +030088 u64 dma_base_and_ceiling;
Rob Clark16ea9752013-01-08 15:04:28 -060089
Rob Clark16ea9752013-01-08 15:04:28 -060090 gem = drm_fb_cma_get_gem_obj(fb, 0);
91
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030092 start = gem->paddr + fb->offsets[0] +
93 crtc->y * fb->pitches[0] +
Laurent Pinchart59f11a42016-10-18 01:41:14 +030094 crtc->x * drm_format_plane_cpp(fb->pixel_format, 0);
Rob Clark16ea9752013-01-08 15:04:28 -060095
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030096 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
Rob Clark16ea9752013-01-08 15:04:28 -060097
Jyri Sarha7eb9f062016-08-26 15:10:14 +030098 /* Write LCDC_DMA_FB_BASE_ADDR_0_REG and LCDC_DMA_FB_CEILING_ADDR_0_REG
99 * with a single insruction, if available. This should make it more
100 * unlikely that LCDC would fetch the DMA addresses in the middle of
101 * an update.
102 */
Daniel Schultz4c268d62016-10-28 13:52:41 +0200103 if (priv->rev == 1)
104 end -= 1;
105
106 dma_base_and_ceiling = (u64)end << 32 | start;
Jyri Sarha7eb9f062016-08-26 15:10:14 +0300107 tilcdc_write64(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_base_and_ceiling);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300108
109 if (tilcdc_crtc->curr_fb)
110 drm_flip_work_queue(&tilcdc_crtc->unref_work,
111 tilcdc_crtc->curr_fb);
112
113 tilcdc_crtc->curr_fb = fb;
Rob Clark16ea9752013-01-08 15:04:28 -0600114}
115
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100116/*
117 * The driver currently only supports the RGB565 format for revision 1. For
118 * 16 bits-per-pixel the palette block is bypassed, but the first 32 bytes of
119 * the framebuffer are still considered palette. The first 16-bit entry must
120 * be 0x4000 while all other entries must be zeroed.
121 */
122static void tilcdc_crtc_load_palette(struct drm_crtc *crtc)
123{
124 u32 dma_fb_base, dma_fb_ceiling, raster_ctl;
125 struct tilcdc_crtc *tilcdc_crtc;
126 struct drm_device *dev;
127 u16 *first_entry;
128
129 dev = crtc->dev;
130 tilcdc_crtc = to_tilcdc_crtc(crtc);
131 first_entry = tilcdc_crtc->palette_base;
132
133 *first_entry = TILCDC_REV1_PALETTE_FIRST_ENTRY;
134
135 dma_fb_base = tilcdc_read(dev, LCDC_DMA_FB_BASE_ADDR_0_REG);
136 dma_fb_ceiling = tilcdc_read(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG);
137 raster_ctl = tilcdc_read(dev, LCDC_RASTER_CTRL_REG);
138
139 /* Tell the LCDC where the palette is located. */
140 tilcdc_write(dev, LCDC_DMA_FB_BASE_ADDR_0_REG,
141 tilcdc_crtc->palette_dma_handle);
142 tilcdc_write(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG,
143 (u32)tilcdc_crtc->palette_dma_handle
144 + TILCDC_REV1_PALETTE_SIZE - 1);
145
146 /* Load it. */
147 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
148 LCDC_PALETTE_LOAD_MODE(DATA_ONLY));
149 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
150 LCDC_PALETTE_LOAD_MODE(PALETTE_ONLY));
151
152 /* Enable the LCDC and wait for palette to be loaded. */
153 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
154 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
155
156 wait_for_completion(&tilcdc_crtc->palette_loaded);
157
158 /* Restore the registers. */
159 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
160 tilcdc_write(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_fb_base);
161 tilcdc_write(dev, LCDC_DMA_FB_CEILING_ADDR_0_REG, dma_fb_ceiling);
162 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, raster_ctl);
163}
164
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300165static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
166{
167 struct tilcdc_drm_private *priv = dev->dev_private;
168
169 tilcdc_clear_irqstatus(dev, 0xffffffff);
170
171 if (priv->rev == 1) {
172 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarhacba88442016-11-16 00:12:27 +0200173 LCDC_V1_SYNC_LOST_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300174 LCDC_V1_UNDERFLOW_INT_ENA);
Karl Beldan8d6c3f72016-08-23 12:57:00 +0000175 tilcdc_set(dev, LCDC_DMA_CTRL_REG,
176 LCDC_V1_END_OF_FRAME_INT_ENA);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300177 } else {
178 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
179 LCDC_V2_UNDERFLOW_INT_ENA |
180 LCDC_V2_END_OF_FRAME0_INT_ENA |
181 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
182 }
183}
184
185static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
186{
187 struct tilcdc_drm_private *priv = dev->dev_private;
188
189 /* disable irqs that we might have enabled: */
190 if (priv->rev == 1) {
191 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarhacba88442016-11-16 00:12:27 +0200192 LCDC_V1_SYNC_LOST_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300193 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
194 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
195 LCDC_V1_END_OF_FRAME_INT_ENA);
196 } else {
197 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
198 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
199 LCDC_V2_END_OF_FRAME0_INT_ENA |
200 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
201 }
202}
203
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300204static void reset(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600205{
206 struct drm_device *dev = crtc->dev;
207 struct tilcdc_drm_private *priv = dev->dev_private;
208
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300209 if (priv->rev != 2)
210 return;
211
212 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
213 usleep_range(250, 1000);
214 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
215}
216
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300217static void tilcdc_crtc_enable(struct drm_crtc *crtc)
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300218{
219 struct drm_device *dev = crtc->dev;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300220 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100221 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300222
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300223 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300224 mutex_lock(&tilcdc_crtc->enable_lock);
225 if (tilcdc_crtc->enabled || tilcdc_crtc->shutdown) {
226 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300227 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300228 }
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300229
230 pm_runtime_get_sync(dev->dev);
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300231
232 reset(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600233
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100234 if (priv->rev == 1 && !completion_done(&tilcdc_crtc->palette_loaded))
235 tilcdc_crtc_load_palette(crtc);
236
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300237 tilcdc_crtc_enable_irqs(dev);
238
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300239 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
Rob Clark16ea9752013-01-08 15:04:28 -0600240 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_PALETTE_LOAD_MODE(DATA_ONLY));
241 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300242
243 drm_crtc_vblank_on(crtc);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300244
245 tilcdc_crtc->enabled = true;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300246 mutex_unlock(&tilcdc_crtc->enable_lock);
Rob Clark16ea9752013-01-08 15:04:28 -0600247}
248
Jyri Sarha2d53a182016-10-25 12:27:31 +0300249static void tilcdc_crtc_off(struct drm_crtc *crtc, bool shutdown)
Rob Clark16ea9752013-01-08 15:04:28 -0600250{
Jyri Sarha2d5be882016-04-07 20:20:23 +0300251 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600252 struct drm_device *dev = crtc->dev;
Jyri Sarha2d5be882016-04-07 20:20:23 +0300253 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600254
Jyri Sarha2d53a182016-10-25 12:27:31 +0300255 mutex_lock(&tilcdc_crtc->enable_lock);
256 if (shutdown)
257 tilcdc_crtc->shutdown = true;
258 if (!tilcdc_crtc->enabled) {
259 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300260 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300261 }
Jyri Sarha2d5be882016-04-07 20:20:23 +0300262 tilcdc_crtc->frame_done = false;
Rob Clark16ea9752013-01-08 15:04:28 -0600263 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha2d5be882016-04-07 20:20:23 +0300264
265 /*
266 * if necessary wait for framedone irq which will still come
267 * before putting things to sleep..
268 */
269 if (priv->rev == 2) {
270 int ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
271 tilcdc_crtc->frame_done,
Jyri Sarha437c7d92016-06-16 16:19:17 +0300272 msecs_to_jiffies(500));
Jyri Sarha2d5be882016-04-07 20:20:23 +0300273 if (ret == 0)
274 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
275 __func__);
276 }
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300277
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100278 /*
279 * LCDC will not retain the palette when reset. Make sure it gets
280 * reloaded on tilcdc_crtc_enable().
281 */
282 if (priv->rev == 1)
283 reinit_completion(&tilcdc_crtc->palette_loaded);
284
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300285 drm_crtc_vblank_off(crtc);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300286
287 tilcdc_crtc_disable_irqs(dev);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300288
289 pm_runtime_put_sync(dev->dev);
290
291 if (tilcdc_crtc->next_fb) {
292 drm_flip_work_queue(&tilcdc_crtc->unref_work,
293 tilcdc_crtc->next_fb);
294 tilcdc_crtc->next_fb = NULL;
295 }
296
297 if (tilcdc_crtc->curr_fb) {
298 drm_flip_work_queue(&tilcdc_crtc->unref_work,
299 tilcdc_crtc->curr_fb);
300 tilcdc_crtc->curr_fb = NULL;
301 }
302
303 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
304 tilcdc_crtc->last_vblank = ktime_set(0, 0);
305
306 tilcdc_crtc->enabled = false;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300307 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300308}
309
Jyri Sarha9e79e062016-10-18 23:23:27 +0300310static void tilcdc_crtc_disable(struct drm_crtc *crtc)
311{
312 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300313 tilcdc_crtc_off(crtc, false);
314}
315
316void tilcdc_crtc_shutdown(struct drm_crtc *crtc)
317{
318 tilcdc_crtc_off(crtc, true);
Jyri Sarha9e79e062016-10-18 23:23:27 +0300319}
320
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300321static bool tilcdc_crtc_is_on(struct drm_crtc *crtc)
322{
323 return crtc->state && crtc->state->enable && crtc->state->active;
Rob Clark16ea9752013-01-08 15:04:28 -0600324}
325
Jyri Sarha13b3d722016-04-06 14:02:38 +0300326static void tilcdc_crtc_recover_work(struct work_struct *work)
327{
328 struct tilcdc_crtc *tilcdc_crtc =
329 container_of(work, struct tilcdc_crtc, recover_work);
330 struct drm_crtc *crtc = &tilcdc_crtc->base;
331
332 dev_info(crtc->dev->dev, "%s: Reset CRTC", __func__);
333
334 drm_modeset_lock_crtc(crtc, NULL);
335
336 if (!tilcdc_crtc_is_on(crtc))
337 goto out;
338
339 tilcdc_crtc_disable(crtc);
340 tilcdc_crtc_enable(crtc);
341out:
342 drm_modeset_unlock_crtc(crtc);
343}
344
Rob Clark16ea9752013-01-08 15:04:28 -0600345static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
346{
347 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Jyri Sarha4e910c72016-09-06 22:55:33 +0300348 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600349
Jyri Sarha6c94c712016-09-07 11:46:40 +0300350 drm_modeset_lock_crtc(crtc, NULL);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300351 tilcdc_crtc_disable(crtc);
Jyri Sarha6c94c712016-09-07 11:46:40 +0300352 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600353
Jyri Sarha4e910c72016-09-06 22:55:33 +0300354 flush_workqueue(priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600355
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300356 of_node_put(crtc->port);
Rob Clark16ea9752013-01-08 15:04:28 -0600357 drm_crtc_cleanup(crtc);
Rob Clarka464d612013-08-07 13:41:20 -0400358 drm_flip_work_cleanup(&tilcdc_crtc->unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -0600359}
360
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300361int tilcdc_crtc_update_fb(struct drm_crtc *crtc,
Rob Clark16ea9752013-01-08 15:04:28 -0600362 struct drm_framebuffer *fb,
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300363 struct drm_pending_vblank_event *event)
Rob Clark16ea9752013-01-08 15:04:28 -0600364{
365 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
366 struct drm_device *dev = crtc->dev;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300367 unsigned long flags;
Tomi Valkeinen6f206e92014-02-07 17:37:07 +0000368
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300369 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
370
Rob Clark16ea9752013-01-08 15:04:28 -0600371 if (tilcdc_crtc->event) {
372 dev_err(dev->dev, "already pending page flip!\n");
373 return -EBUSY;
374 }
375
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300376 drm_framebuffer_reference(fb);
377
Matt Roperf4510a22014-04-01 15:22:40 -0700378 crtc->primary->fb = fb;
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300379
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200380 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300381
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300382 if (crtc->hwmode.vrefresh && ktime_to_ns(tilcdc_crtc->last_vblank)) {
383 ktime_t next_vblank;
384 s64 tdiff;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300385
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300386 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
387 1000000 / crtc->hwmode.vrefresh);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200388
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300389 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
390
391 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
392 tilcdc_crtc->next_fb = fb;
393 }
394
395 if (tilcdc_crtc->next_fb != fb)
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200396 set_scanout(crtc, fb);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200397
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300398 tilcdc_crtc->event = event;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200399
400 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600401
402 return 0;
403}
404
Rob Clark16ea9752013-01-08 15:04:28 -0600405static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
406 const struct drm_display_mode *mode,
407 struct drm_display_mode *adjusted_mode)
408{
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200409 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
410
411 if (!tilcdc_crtc->simulate_vesa_sync)
412 return true;
413
414 /*
415 * tilcdc does not generate VESA-compliant sync but aligns
416 * VS on the second edge of HS instead of first edge.
417 * We use adjusted_mode, to fixup sync by aligning both rising
418 * edges and add HSKEW offset to fix the sync.
419 */
420 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
421 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
422
423 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
424 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
425 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
426 } else {
427 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
428 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
429 }
430
Rob Clark16ea9752013-01-08 15:04:28 -0600431 return true;
432}
433
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200434/*
435 * Calculate the percentage difference between the requested pixel clock rate
436 * and the effective rate resulting from calculating the clock divider value.
437 */
438static unsigned int tilcdc_pclk_diff(unsigned long rate,
439 unsigned long real_rate)
440{
441 int r = rate / 100, rr = real_rate / 100;
442
443 return (unsigned int)(abs(((rr - r) * 100) / r));
444}
445
Jyri Sarha642e5162016-09-06 16:19:54 +0300446static void tilcdc_crtc_set_clk(struct drm_crtc *crtc)
447{
448 struct drm_device *dev = crtc->dev;
449 struct tilcdc_drm_private *priv = dev->dev_private;
450 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200451 unsigned long clk_rate, real_rate, req_rate;
452 unsigned int clkdiv;
Jyri Sarha642e5162016-09-06 16:19:54 +0300453 int ret;
454
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200455 clkdiv = 2; /* first try using a standard divider of 2 */
456
Jyri Sarha642e5162016-09-06 16:19:54 +0300457 /* mode.clock is in KHz, set_rate wants parameter in Hz */
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200458 req_rate = crtc->mode.clock * 1000;
459
460 ret = clk_set_rate(priv->clk, req_rate * clkdiv);
461 clk_rate = clk_get_rate(priv->clk);
Jyri Sarha642e5162016-09-06 16:19:54 +0300462 if (ret < 0) {
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200463 /*
464 * If we fail to set the clock rate (some architectures don't
465 * use the common clock framework yet and may not implement
466 * all the clk API calls for every clock), try the next best
467 * thing: adjusting the clock divider, unless clk_get_rate()
468 * failed as well.
469 */
470 if (!clk_rate) {
471 /* Nothing more we can do. Just bail out. */
472 dev_err(dev->dev,
473 "failed to set the pixel clock - unable to read current lcdc clock rate\n");
474 return;
475 }
476
477 clkdiv = DIV_ROUND_CLOSEST(clk_rate, req_rate);
478
479 /*
480 * Emit a warning if the real clock rate resulting from the
481 * calculated divider differs much from the requested rate.
482 *
483 * 5% is an arbitrary value - LCDs are usually quite tolerant
484 * about pixel clock rates.
485 */
486 real_rate = clkdiv * req_rate;
487
488 if (tilcdc_pclk_diff(clk_rate, real_rate) > 5) {
489 dev_warn(dev->dev,
490 "effective pixel clock rate (%luHz) differs from the calculated rate (%luHz)\n",
491 clk_rate, real_rate);
492 }
Jyri Sarha642e5162016-09-06 16:19:54 +0300493 }
494
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200495 tilcdc_crtc->lcd_fck_rate = clk_rate;
Jyri Sarha642e5162016-09-06 16:19:54 +0300496
497 DBG("lcd_clk=%u, mode clock=%d, div=%u",
498 tilcdc_crtc->lcd_fck_rate, crtc->mode.clock, clkdiv);
499
500 /* Configure the LCD clock divisor. */
501 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
502 LCDC_RASTER_MODE);
503
504 if (priv->rev == 2)
505 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
506 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
507 LCDC_V2_CORE_CLK_EN);
508}
509
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300510static void tilcdc_crtc_mode_set_nofb(struct drm_crtc *crtc)
511{
512 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
513 struct drm_device *dev = crtc->dev;
514 struct tilcdc_drm_private *priv = dev->dev_private;
515 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
516 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
517 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
518 struct drm_framebuffer *fb = crtc->primary->state->fb;
519
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300520 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
521
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300522 if (WARN_ON(!info))
523 return;
524
525 if (WARN_ON(!fb))
526 return;
527
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300528 /* Configure the Burst Size and fifo threshold of DMA: */
529 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
530 switch (info->dma_burst_sz) {
531 case 1:
532 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
533 break;
534 case 2:
535 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
536 break;
537 case 4:
538 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
539 break;
540 case 8:
541 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
542 break;
543 case 16:
544 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
545 break;
546 default:
547 dev_err(dev->dev, "invalid burst size\n");
548 return;
549 }
550 reg |= (info->fifo_th << 8);
551 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
552
553 /* Configure timings: */
554 hbp = mode->htotal - mode->hsync_end;
555 hfp = mode->hsync_start - mode->hdisplay;
556 hsw = mode->hsync_end - mode->hsync_start;
557 vbp = mode->vtotal - mode->vsync_end;
558 vfp = mode->vsync_start - mode->vdisplay;
559 vsw = mode->vsync_end - mode->vsync_start;
560
561 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
562 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
563
564 /* Set AC Bias Period and Number of Transitions per Interrupt: */
565 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
566 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
567 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
568
569 /*
570 * subtract one from hfp, hbp, hsw because the hardware uses
571 * a value of 0 as 1
572 */
573 if (priv->rev == 2) {
574 /* clear bits we're going to set */
575 reg &= ~0x78000033;
576 reg |= ((hfp-1) & 0x300) >> 8;
577 reg |= ((hbp-1) & 0x300) >> 4;
578 reg |= ((hsw-1) & 0x3c0) << 21;
579 }
580 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
581
582 reg = (((mode->hdisplay >> 4) - 1) << 4) |
583 (((hbp-1) & 0xff) << 24) |
584 (((hfp-1) & 0xff) << 16) |
585 (((hsw-1) & 0x3f) << 10);
586 if (priv->rev == 2)
587 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
588 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
589
590 reg = ((mode->vdisplay - 1) & 0x3ff) |
591 ((vbp & 0xff) << 24) |
592 ((vfp & 0xff) << 16) |
593 (((vsw-1) & 0x3f) << 10);
594 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
595
596 /*
597 * be sure to set Bit 10 for the V2 LCDC controller,
598 * otherwise limited to 1024 pixels width, stopping
599 * 1920x1080 being supported.
600 */
601 if (priv->rev == 2) {
602 if ((mode->vdisplay - 1) & 0x400) {
603 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
604 LCDC_LPP_B10);
605 } else {
606 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
607 LCDC_LPP_B10);
608 }
609 }
610
611 /* Configure display type: */
612 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
613 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
614 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
615 0x000ff000 /* Palette Loading Delay bits */);
616 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
617 if (info->tft_alt_mode)
618 reg |= LCDC_TFT_ALT_ENABLE;
619 if (priv->rev == 2) {
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300620 switch (fb->pixel_format) {
621 case DRM_FORMAT_BGR565:
622 case DRM_FORMAT_RGB565:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300623 break;
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300624 case DRM_FORMAT_XBGR8888:
625 case DRM_FORMAT_XRGB8888:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300626 reg |= LCDC_V2_TFT_24BPP_UNPACK;
627 /* fallthrough */
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300628 case DRM_FORMAT_BGR888:
629 case DRM_FORMAT_RGB888:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300630 reg |= LCDC_V2_TFT_24BPP_MODE;
631 break;
632 default:
633 dev_err(dev->dev, "invalid pixel format\n");
634 return;
635 }
636 }
637 reg |= info->fdd < 12;
638 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
639
640 if (info->invert_pxl_clk)
641 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
642 else
643 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
644
645 if (info->sync_ctrl)
646 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
647 else
648 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
649
650 if (info->sync_edge)
651 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
652 else
653 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
654
655 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
656 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
657 else
658 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
659
660 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
661 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
662 else
663 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
664
665 if (info->raster_order)
666 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
667 else
668 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
669
670 drm_framebuffer_reference(fb);
671
672 set_scanout(crtc, fb);
673
Jyri Sarha642e5162016-09-06 16:19:54 +0300674 tilcdc_crtc_set_clk(crtc);
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300675
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300676 crtc->hwmode = crtc->state->adjusted_mode;
677}
678
Jyri Sarhadb380c52016-04-07 15:10:23 +0300679static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
680 struct drm_crtc_state *state)
681{
682 struct drm_display_mode *mode = &state->mode;
683 int ret;
684
685 /* If we are not active we don't care */
686 if (!state->active)
687 return 0;
688
689 if (state->state->planes[0].ptr != crtc->primary ||
690 state->state->planes[0].state == NULL ||
691 state->state->planes[0].state->crtc != crtc) {
692 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
693 return -EINVAL;
694 }
695
696 ret = tilcdc_crtc_mode_valid(crtc, mode);
697 if (ret) {
698 dev_dbg(crtc->dev->dev, "Mode \"%s\" not valid", mode->name);
699 return -EINVAL;
700 }
701
702 return 0;
703}
704
Rob Clark16ea9752013-01-08 15:04:28 -0600705static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
Jyri Sarha305198d2016-04-07 15:05:16 +0300706 .destroy = tilcdc_crtc_destroy,
707 .set_config = drm_atomic_helper_set_config,
708 .page_flip = drm_atomic_helper_page_flip,
709 .reset = drm_atomic_helper_crtc_reset,
710 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
711 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
Rob Clark16ea9752013-01-08 15:04:28 -0600712};
713
714static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
Rob Clark16ea9752013-01-08 15:04:28 -0600715 .mode_fixup = tilcdc_crtc_mode_fixup,
Jyri Sarha305198d2016-04-07 15:05:16 +0300716 .enable = tilcdc_crtc_enable,
717 .disable = tilcdc_crtc_disable,
Jyri Sarhadb380c52016-04-07 15:10:23 +0300718 .atomic_check = tilcdc_crtc_atomic_check,
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300719 .mode_set_nofb = tilcdc_crtc_mode_set_nofb,
Rob Clark16ea9752013-01-08 15:04:28 -0600720};
721
722int tilcdc_crtc_max_width(struct drm_crtc *crtc)
723{
724 struct drm_device *dev = crtc->dev;
725 struct tilcdc_drm_private *priv = dev->dev_private;
726 int max_width = 0;
727
728 if (priv->rev == 1)
729 max_width = 1024;
730 else if (priv->rev == 2)
731 max_width = 2048;
732
733 return max_width;
734}
735
736int tilcdc_crtc_mode_valid(struct drm_crtc *crtc, struct drm_display_mode *mode)
737{
738 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
739 unsigned int bandwidth;
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500740 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
Rob Clark16ea9752013-01-08 15:04:28 -0600741
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500742 /*
743 * check to see if the width is within the range that
744 * the LCD Controller physically supports
745 */
Rob Clark16ea9752013-01-08 15:04:28 -0600746 if (mode->hdisplay > tilcdc_crtc_max_width(crtc))
747 return MODE_VIRTUAL_X;
748
749 /* width must be multiple of 16 */
750 if (mode->hdisplay & 0xf)
751 return MODE_VIRTUAL_X;
752
753 if (mode->vdisplay > 2048)
754 return MODE_VIRTUAL_Y;
755
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500756 DBG("Processing mode %dx%d@%d with pixel clock %d",
757 mode->hdisplay, mode->vdisplay,
758 drm_mode_vrefresh(mode), mode->clock);
759
760 hbp = mode->htotal - mode->hsync_end;
761 hfp = mode->hsync_start - mode->hdisplay;
762 hsw = mode->hsync_end - mode->hsync_start;
763 vbp = mode->vtotal - mode->vsync_end;
764 vfp = mode->vsync_start - mode->vdisplay;
765 vsw = mode->vsync_end - mode->vsync_start;
766
767 if ((hbp-1) & ~0x3ff) {
768 DBG("Pruning mode: Horizontal Back Porch out of range");
769 return MODE_HBLANK_WIDE;
770 }
771
772 if ((hfp-1) & ~0x3ff) {
773 DBG("Pruning mode: Horizontal Front Porch out of range");
774 return MODE_HBLANK_WIDE;
775 }
776
777 if ((hsw-1) & ~0x3ff) {
778 DBG("Pruning mode: Horizontal Sync Width out of range");
779 return MODE_HSYNC_WIDE;
780 }
781
782 if (vbp & ~0xff) {
783 DBG("Pruning mode: Vertical Back Porch out of range");
784 return MODE_VBLANK_WIDE;
785 }
786
787 if (vfp & ~0xff) {
788 DBG("Pruning mode: Vertical Front Porch out of range");
789 return MODE_VBLANK_WIDE;
790 }
791
792 if ((vsw-1) & ~0x3f) {
793 DBG("Pruning mode: Vertical Sync Width out of range");
794 return MODE_VSYNC_WIDE;
795 }
796
Darren Etheridge4e564342013-06-21 13:52:23 -0500797 /*
798 * some devices have a maximum allowed pixel clock
799 * configured from the DT
800 */
801 if (mode->clock > priv->max_pixelclock) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500802 DBG("Pruning mode: pixel clock too high");
Darren Etheridge4e564342013-06-21 13:52:23 -0500803 return MODE_CLOCK_HIGH;
804 }
805
806 /*
807 * some devices further limit the max horizontal resolution
808 * configured from the DT
809 */
810 if (mode->hdisplay > priv->max_width)
811 return MODE_BAD_WIDTH;
812
Rob Clark16ea9752013-01-08 15:04:28 -0600813 /* filter out modes that would require too much memory bandwidth: */
Darren Etheridge4e564342013-06-21 13:52:23 -0500814 bandwidth = mode->hdisplay * mode->vdisplay *
815 drm_mode_vrefresh(mode);
816 if (bandwidth > priv->max_bandwidth) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500817 DBG("Pruning mode: exceeds defined bandwidth limit");
Rob Clark16ea9752013-01-08 15:04:28 -0600818 return MODE_BAD;
Darren Etheridge4e564342013-06-21 13:52:23 -0500819 }
Rob Clark16ea9752013-01-08 15:04:28 -0600820
821 return MODE_OK;
822}
823
824void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
825 const struct tilcdc_panel_info *info)
826{
827 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
828 tilcdc_crtc->info = info;
829}
830
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200831void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
832 bool simulate_vesa_sync)
833{
834 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
835
836 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
837}
838
Rob Clark16ea9752013-01-08 15:04:28 -0600839void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
840{
Rob Clark16ea9752013-01-08 15:04:28 -0600841 struct drm_device *dev = crtc->dev;
842 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarha642e5162016-09-06 16:19:54 +0300843 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600844
Jyri Sarha642e5162016-09-06 16:19:54 +0300845 drm_modeset_lock_crtc(crtc, NULL);
846 if (tilcdc_crtc->lcd_fck_rate != clk_get_rate(priv->clk)) {
847 if (tilcdc_crtc_is_on(crtc)) {
848 pm_runtime_get_sync(dev->dev);
849 tilcdc_crtc_disable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600850
Jyri Sarha642e5162016-09-06 16:19:54 +0300851 tilcdc_crtc_set_clk(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600852
Jyri Sarha642e5162016-09-06 16:19:54 +0300853 tilcdc_crtc_enable(crtc);
854 pm_runtime_put_sync(dev->dev);
855 }
Rob Clark16ea9752013-01-08 15:04:28 -0600856 }
Jyri Sarha642e5162016-09-06 16:19:54 +0300857 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600858}
859
Jyri Sarha5895d082016-01-08 14:33:09 +0200860#define SYNC_LOST_COUNT_LIMIT 50
861
Rob Clark16ea9752013-01-08 15:04:28 -0600862irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
863{
864 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
865 struct drm_device *dev = crtc->dev;
866 struct tilcdc_drm_private *priv = dev->dev_private;
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300867 uint32_t stat;
Rob Clark16ea9752013-01-08 15:04:28 -0600868
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300869 stat = tilcdc_read_irqstatus(dev);
870 tilcdc_clear_irqstatus(dev, stat);
871
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300872 if (stat & LCDC_END_OF_FRAME0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600873 unsigned long flags;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200874 bool skip_event = false;
875 ktime_t now;
876
877 now = ktime_get();
Rob Clark16ea9752013-01-08 15:04:28 -0600878
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300879 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600880
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200881 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600882
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200883 tilcdc_crtc->last_vblank = now;
Rob Clark16ea9752013-01-08 15:04:28 -0600884
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200885 if (tilcdc_crtc->next_fb) {
886 set_scanout(crtc, tilcdc_crtc->next_fb);
887 tilcdc_crtc->next_fb = NULL;
888 skip_event = true;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300889 }
890
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200891 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
892
Gustavo Padovan099ede82016-07-04 21:04:52 -0300893 drm_crtc_handle_vblank(crtc);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200894
895 if (!skip_event) {
896 struct drm_pending_vblank_event *event;
897
898 spin_lock_irqsave(&dev->event_lock, flags);
899
900 event = tilcdc_crtc->event;
901 tilcdc_crtc->event = NULL;
902 if (event)
Gustavo Padovandfebc152016-04-14 10:48:22 -0700903 drm_crtc_send_vblank_event(crtc, event);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200904
905 spin_unlock_irqrestore(&dev->event_lock, flags);
906 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200907
908 if (tilcdc_crtc->frame_intact)
909 tilcdc_crtc->sync_lost_count = 0;
910 else
911 tilcdc_crtc->frame_intact = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600912 }
913
Jyri Sarha14944112016-04-07 20:36:48 +0300914 if (stat & LCDC_FIFO_UNDERFLOW)
Daniel Schultzd7014532016-10-28 13:52:42 +0200915 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underflow",
Jyri Sarha14944112016-04-07 20:36:48 +0300916 __func__, stat);
917
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100918 if (priv->rev == 1) {
919 if (stat & LCDC_PL_LOAD_DONE) {
920 complete(&tilcdc_crtc->palette_loaded);
921 tilcdc_clear(dev,
922 LCDC_RASTER_CTRL_REG, LCDC_V1_PL_INT_ENA);
923 }
924 }
925
Jyri Sarhacba88442016-11-16 00:12:27 +0200926 if (stat & LCDC_SYNC_LOST) {
927 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
928 __func__, stat);
929 tilcdc_crtc->frame_intact = false;
930 if (tilcdc_crtc->sync_lost_count++ >
931 SYNC_LOST_COUNT_LIMIT) {
932 dev_err(dev->dev, "%s(0x%08x): Sync lost flood detected, recovering", __func__, stat);
933 queue_work(system_wq, &tilcdc_crtc->recover_work);
934 if (priv->rev == 1)
935 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
936 LCDC_V1_SYNC_LOST_INT_ENA);
937 else
938 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
939 LCDC_SYNC_LOST);
940 tilcdc_crtc->sync_lost_count = 0;
941 }
942 }
943
Jyri Sarha14944112016-04-07 20:36:48 +0300944 /* For revision 2 only */
Rob Clark16ea9752013-01-08 15:04:28 -0600945 if (priv->rev == 2) {
946 if (stat & LCDC_FRAME_DONE) {
947 tilcdc_crtc->frame_done = true;
948 wake_up(&tilcdc_crtc->frame_done_wq);
949 }
Rob Clark16ea9752013-01-08 15:04:28 -0600950
Jyri Sarha14944112016-04-07 20:36:48 +0300951 /* Indicate to LCDC that the interrupt service routine has
952 * completed, see 13.3.6.1.6 in AM335x TRM.
953 */
954 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
955 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200956
Rob Clark16ea9752013-01-08 15:04:28 -0600957 return IRQ_HANDLED;
958}
959
Rob Clark16ea9752013-01-08 15:04:28 -0600960struct drm_crtc *tilcdc_crtc_create(struct drm_device *dev)
961{
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300962 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600963 struct tilcdc_crtc *tilcdc_crtc;
964 struct drm_crtc *crtc;
965 int ret;
966
Jyri Sarhad0ec32c2016-02-23 12:44:27 +0200967 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
Rob Clark16ea9752013-01-08 15:04:28 -0600968 if (!tilcdc_crtc) {
969 dev_err(dev->dev, "allocation failed\n");
970 return NULL;
971 }
972
Bartosz Golaszewski93452352016-10-31 15:19:26 +0100973 if (priv->rev == 1) {
974 init_completion(&tilcdc_crtc->palette_loaded);
975 tilcdc_crtc->palette_base = dmam_alloc_coherent(dev->dev,
976 TILCDC_REV1_PALETTE_SIZE,
977 &tilcdc_crtc->palette_dma_handle,
978 GFP_KERNEL | __GFP_ZERO);
979 if (!tilcdc_crtc->palette_base)
980 return ERR_PTR(-ENOMEM);
981 }
982
Rob Clark16ea9752013-01-08 15:04:28 -0600983 crtc = &tilcdc_crtc->base;
984
Jyri Sarha47f571c2016-04-07 15:04:18 +0300985 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
986 if (ret < 0)
987 goto fail;
988
Jyri Sarha2d53a182016-10-25 12:27:31 +0300989 mutex_init(&tilcdc_crtc->enable_lock);
990
Rob Clark16ea9752013-01-08 15:04:28 -0600991 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
992
Boris BREZILLONd7f8db52014-11-14 19:30:30 +0100993 drm_flip_work_init(&tilcdc_crtc->unref_work,
Rob Clarka464d612013-08-07 13:41:20 -0400994 "unref", unref_worker);
Rob Clark16ea9752013-01-08 15:04:28 -0600995
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200996 spin_lock_init(&tilcdc_crtc->irq_lock);
Jyri Sarha13b3d722016-04-06 14:02:38 +0300997 INIT_WORK(&tilcdc_crtc->recover_work, tilcdc_crtc_recover_work);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200998
Jyri Sarha47f571c2016-04-07 15:04:18 +0300999 ret = drm_crtc_init_with_planes(dev, crtc,
1000 &tilcdc_crtc->primary,
1001 NULL,
1002 &tilcdc_crtc_funcs,
1003 "tilcdc crtc");
Rob Clark16ea9752013-01-08 15:04:28 -06001004 if (ret < 0)
1005 goto fail;
1006
1007 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
1008
Jyri Sarhad66284fb2015-05-27 11:58:37 +03001009 if (priv->is_componentized) {
1010 struct device_node *ports =
1011 of_get_child_by_name(dev->dev->of_node, "ports");
1012
1013 if (ports) {
1014 crtc->port = of_get_child_by_name(ports, "port");
1015 of_node_put(ports);
1016 } else {
1017 crtc->port =
1018 of_get_child_by_name(dev->dev->of_node, "port");
1019 }
1020 if (!crtc->port) { /* This should never happen */
1021 dev_err(dev->dev, "Port node not found in %s\n",
1022 dev->dev->of_node->full_name);
1023 goto fail;
1024 }
1025 }
1026
Rob Clark16ea9752013-01-08 15:04:28 -06001027 return crtc;
1028
1029fail:
1030 tilcdc_crtc_destroy(crtc);
1031 return NULL;
1032}