blob: 01692409f61335369b334ed5d68fd6894ad9278f [file] [log] [blame]
Rob Clark16ea9752013-01-08 15:04:28 -06001/*
2 * Copyright (C) 2012 Texas Instruments
3 * Author: Rob Clark <robdclark@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program. If not, see <http://www.gnu.org/licenses/>.
16 */
17
Sean Paulce2f2c32016-09-21 06:14:53 -070018#include <drm/drm_atomic.h>
Jyri Sarha305198d2016-04-07 15:05:16 +030019#include <drm/drm_atomic_helper.h>
Sean Paulce2f2c32016-09-21 06:14:53 -070020#include <drm/drm_crtc.h>
21#include <drm/drm_flip_work.h>
22#include <drm/drm_plane_helper.h>
Jyri Sarha4e910c72016-09-06 22:55:33 +030023#include <linux/workqueue.h>
Rob Clark16ea9752013-01-08 15:04:28 -060024
25#include "tilcdc_drv.h"
26#include "tilcdc_regs.h"
27
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020028#define TILCDC_VBLANK_SAFETY_THRESHOLD_US 1000
29
Rob Clark16ea9752013-01-08 15:04:28 -060030struct tilcdc_crtc {
31 struct drm_crtc base;
32
Jyri Sarha47f571c2016-04-07 15:04:18 +030033 struct drm_plane primary;
Rob Clark16ea9752013-01-08 15:04:28 -060034 const struct tilcdc_panel_info *info;
Rob Clark16ea9752013-01-08 15:04:28 -060035 struct drm_pending_vblank_event *event;
Jyri Sarha2d53a182016-10-25 12:27:31 +030036 struct mutex enable_lock;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +030037 bool enabled;
Jyri Sarha2d53a182016-10-25 12:27:31 +030038 bool shutdown;
Rob Clark16ea9752013-01-08 15:04:28 -060039 wait_queue_head_t frame_done_wq;
40 bool frame_done;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020041 spinlock_t irq_lock;
42
Jyri Sarha642e5162016-09-06 16:19:54 +030043 unsigned int lcd_fck_rate;
44
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020045 ktime_t last_vblank;
Rob Clark16ea9752013-01-08 15:04:28 -060046
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030047 struct drm_framebuffer *curr_fb;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +020048 struct drm_framebuffer *next_fb;
Rob Clark16ea9752013-01-08 15:04:28 -060049
50 /* for deferred fb unref's: */
Rob Clarka464d612013-08-07 13:41:20 -040051 struct drm_flip_work unref_work;
Jyri Sarha103cd8b2015-02-10 14:13:23 +020052
53 /* Only set if an external encoder is connected */
54 bool simulate_vesa_sync;
Jyri Sarha5895d082016-01-08 14:33:09 +020055
56 int sync_lost_count;
57 bool frame_intact;
Jyri Sarha13b3d722016-04-06 14:02:38 +030058 struct work_struct recover_work;
Rob Clark16ea9752013-01-08 15:04:28 -060059};
60#define to_tilcdc_crtc(x) container_of(x, struct tilcdc_crtc, base)
61
Rob Clarka464d612013-08-07 13:41:20 -040062static void unref_worker(struct drm_flip_work *work, void *val)
Rob Clark16ea9752013-01-08 15:04:28 -060063{
Darren Etheridgef7b45752013-06-21 13:52:26 -050064 struct tilcdc_crtc *tilcdc_crtc =
Rob Clarka464d612013-08-07 13:41:20 -040065 container_of(work, struct tilcdc_crtc, unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -060066 struct drm_device *dev = tilcdc_crtc->base.dev;
Rob Clark16ea9752013-01-08 15:04:28 -060067
68 mutex_lock(&dev->mode_config.mutex);
Rob Clarka464d612013-08-07 13:41:20 -040069 drm_framebuffer_unreference(val);
Rob Clark16ea9752013-01-08 15:04:28 -060070 mutex_unlock(&dev->mode_config.mutex);
71}
72
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030073static void set_scanout(struct drm_crtc *crtc, struct drm_framebuffer *fb)
Rob Clark16ea9752013-01-08 15:04:28 -060074{
75 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
76 struct drm_device *dev = crtc->dev;
Daniel Schultz4c268d62016-10-28 13:52:41 +020077 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -060078 struct drm_gem_cma_object *gem;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030079 dma_addr_t start, end;
Jyri Sarha7eb9f062016-08-26 15:10:14 +030080 u64 dma_base_and_ceiling;
Rob Clark16ea9752013-01-08 15:04:28 -060081
Rob Clark16ea9752013-01-08 15:04:28 -060082 gem = drm_fb_cma_get_gem_obj(fb, 0);
83
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030084 start = gem->paddr + fb->offsets[0] +
85 crtc->y * fb->pitches[0] +
Laurent Pinchart59f11a42016-10-18 01:41:14 +030086 crtc->x * drm_format_plane_cpp(fb->pixel_format, 0);
Rob Clark16ea9752013-01-08 15:04:28 -060087
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +030088 end = start + (crtc->mode.vdisplay * fb->pitches[0]);
Rob Clark16ea9752013-01-08 15:04:28 -060089
Jyri Sarha7eb9f062016-08-26 15:10:14 +030090 /* Write LCDC_DMA_FB_BASE_ADDR_0_REG and LCDC_DMA_FB_CEILING_ADDR_0_REG
91 * with a single insruction, if available. This should make it more
92 * unlikely that LCDC would fetch the DMA addresses in the middle of
93 * an update.
94 */
Daniel Schultz4c268d62016-10-28 13:52:41 +020095 if (priv->rev == 1)
96 end -= 1;
97
98 dma_base_and_ceiling = (u64)end << 32 | start;
Jyri Sarha7eb9f062016-08-26 15:10:14 +030099 tilcdc_write64(dev, LCDC_DMA_FB_BASE_ADDR_0_REG, dma_base_and_ceiling);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300100
101 if (tilcdc_crtc->curr_fb)
102 drm_flip_work_queue(&tilcdc_crtc->unref_work,
103 tilcdc_crtc->curr_fb);
104
105 tilcdc_crtc->curr_fb = fb;
Rob Clark16ea9752013-01-08 15:04:28 -0600106}
107
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300108static void tilcdc_crtc_enable_irqs(struct drm_device *dev)
109{
110 struct tilcdc_drm_private *priv = dev->dev_private;
111
112 tilcdc_clear_irqstatus(dev, 0xffffffff);
113
114 if (priv->rev == 1) {
115 tilcdc_set(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarhacba88442016-11-16 00:12:27 +0200116 LCDC_V1_SYNC_LOST_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300117 LCDC_V1_UNDERFLOW_INT_ENA);
Karl Beldan8d6c3f72016-08-23 12:57:00 +0000118 tilcdc_set(dev, LCDC_DMA_CTRL_REG,
119 LCDC_V1_END_OF_FRAME_INT_ENA);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300120 } else {
121 tilcdc_write(dev, LCDC_INT_ENABLE_SET_REG,
122 LCDC_V2_UNDERFLOW_INT_ENA |
123 LCDC_V2_END_OF_FRAME0_INT_ENA |
124 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
125 }
126}
127
128static void tilcdc_crtc_disable_irqs(struct drm_device *dev)
129{
130 struct tilcdc_drm_private *priv = dev->dev_private;
131
132 /* disable irqs that we might have enabled: */
133 if (priv->rev == 1) {
134 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
Jyri Sarhacba88442016-11-16 00:12:27 +0200135 LCDC_V1_SYNC_LOST_INT_ENA |
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300136 LCDC_V1_UNDERFLOW_INT_ENA | LCDC_V1_PL_INT_ENA);
137 tilcdc_clear(dev, LCDC_DMA_CTRL_REG,
138 LCDC_V1_END_OF_FRAME_INT_ENA);
139 } else {
140 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
141 LCDC_V2_UNDERFLOW_INT_ENA | LCDC_V2_PL_INT_ENA |
142 LCDC_V2_END_OF_FRAME0_INT_ENA |
143 LCDC_FRAME_DONE | LCDC_SYNC_LOST);
144 }
145}
146
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300147static void reset(struct drm_crtc *crtc)
Rob Clark16ea9752013-01-08 15:04:28 -0600148{
149 struct drm_device *dev = crtc->dev;
150 struct tilcdc_drm_private *priv = dev->dev_private;
151
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300152 if (priv->rev != 2)
153 return;
154
155 tilcdc_set(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
156 usleep_range(250, 1000);
157 tilcdc_clear(dev, LCDC_CLK_RESET_REG, LCDC_CLK_MAIN_RESET);
158}
159
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300160static void tilcdc_crtc_enable(struct drm_crtc *crtc)
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300161{
162 struct drm_device *dev = crtc->dev;
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300163 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
164
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300165 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300166 mutex_lock(&tilcdc_crtc->enable_lock);
167 if (tilcdc_crtc->enabled || tilcdc_crtc->shutdown) {
168 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300169 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300170 }
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300171
172 pm_runtime_get_sync(dev->dev);
Tomi Valkeinen2efec4f2015-10-20 09:37:27 +0300173
174 reset(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600175
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300176 tilcdc_crtc_enable_irqs(dev);
177
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300178 tilcdc_clear(dev, LCDC_DMA_CTRL_REG, LCDC_DUAL_FRAME_BUFFER_ENABLE);
Rob Clark16ea9752013-01-08 15:04:28 -0600179 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_PALETTE_LOAD_MODE(DATA_ONLY));
180 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300181
182 drm_crtc_vblank_on(crtc);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300183
184 tilcdc_crtc->enabled = true;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300185 mutex_unlock(&tilcdc_crtc->enable_lock);
Rob Clark16ea9752013-01-08 15:04:28 -0600186}
187
Jyri Sarha2d53a182016-10-25 12:27:31 +0300188static void tilcdc_crtc_off(struct drm_crtc *crtc, bool shutdown)
Rob Clark16ea9752013-01-08 15:04:28 -0600189{
Jyri Sarha2d5be882016-04-07 20:20:23 +0300190 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600191 struct drm_device *dev = crtc->dev;
Jyri Sarha2d5be882016-04-07 20:20:23 +0300192 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600193
Jyri Sarha2d53a182016-10-25 12:27:31 +0300194 mutex_lock(&tilcdc_crtc->enable_lock);
195 if (shutdown)
196 tilcdc_crtc->shutdown = true;
197 if (!tilcdc_crtc->enabled) {
198 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300199 return;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300200 }
Jyri Sarha2d5be882016-04-07 20:20:23 +0300201 tilcdc_crtc->frame_done = false;
Rob Clark16ea9752013-01-08 15:04:28 -0600202 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ENABLE);
Jyri Sarha2d5be882016-04-07 20:20:23 +0300203
204 /*
205 * if necessary wait for framedone irq which will still come
206 * before putting things to sleep..
207 */
208 if (priv->rev == 2) {
209 int ret = wait_event_timeout(tilcdc_crtc->frame_done_wq,
210 tilcdc_crtc->frame_done,
Jyri Sarha437c7d92016-06-16 16:19:17 +0300211 msecs_to_jiffies(500));
Jyri Sarha2d5be882016-04-07 20:20:23 +0300212 if (ret == 0)
213 dev_err(dev->dev, "%s: timeout waiting for framedone\n",
214 __func__);
215 }
Jyri Sarhad85f850e2016-06-15 11:16:23 +0300216
217 drm_crtc_vblank_off(crtc);
Jyri Sarhaafaf8332016-06-21 16:00:44 +0300218
219 tilcdc_crtc_disable_irqs(dev);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300220
221 pm_runtime_put_sync(dev->dev);
222
223 if (tilcdc_crtc->next_fb) {
224 drm_flip_work_queue(&tilcdc_crtc->unref_work,
225 tilcdc_crtc->next_fb);
226 tilcdc_crtc->next_fb = NULL;
227 }
228
229 if (tilcdc_crtc->curr_fb) {
230 drm_flip_work_queue(&tilcdc_crtc->unref_work,
231 tilcdc_crtc->curr_fb);
232 tilcdc_crtc->curr_fb = NULL;
233 }
234
235 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
236 tilcdc_crtc->last_vblank = ktime_set(0, 0);
237
238 tilcdc_crtc->enabled = false;
Jyri Sarha2d53a182016-10-25 12:27:31 +0300239 mutex_unlock(&tilcdc_crtc->enable_lock);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300240}
241
Jyri Sarha9e79e062016-10-18 23:23:27 +0300242static void tilcdc_crtc_disable(struct drm_crtc *crtc)
243{
244 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
Jyri Sarha2d53a182016-10-25 12:27:31 +0300245 tilcdc_crtc_off(crtc, false);
246}
247
248void tilcdc_crtc_shutdown(struct drm_crtc *crtc)
249{
250 tilcdc_crtc_off(crtc, true);
Jyri Sarha9e79e062016-10-18 23:23:27 +0300251}
252
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300253static bool tilcdc_crtc_is_on(struct drm_crtc *crtc)
254{
255 return crtc->state && crtc->state->enable && crtc->state->active;
Rob Clark16ea9752013-01-08 15:04:28 -0600256}
257
Jyri Sarha13b3d722016-04-06 14:02:38 +0300258static void tilcdc_crtc_recover_work(struct work_struct *work)
259{
260 struct tilcdc_crtc *tilcdc_crtc =
261 container_of(work, struct tilcdc_crtc, recover_work);
262 struct drm_crtc *crtc = &tilcdc_crtc->base;
263
264 dev_info(crtc->dev->dev, "%s: Reset CRTC", __func__);
265
266 drm_modeset_lock_crtc(crtc, NULL);
267
268 if (!tilcdc_crtc_is_on(crtc))
269 goto out;
270
271 tilcdc_crtc_disable(crtc);
272 tilcdc_crtc_enable(crtc);
273out:
274 drm_modeset_unlock_crtc(crtc);
275}
276
Rob Clark16ea9752013-01-08 15:04:28 -0600277static void tilcdc_crtc_destroy(struct drm_crtc *crtc)
278{
279 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Jyri Sarha4e910c72016-09-06 22:55:33 +0300280 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600281
Jyri Sarha6c94c712016-09-07 11:46:40 +0300282 drm_modeset_lock_crtc(crtc, NULL);
Jyri Sarha47bfd6c2016-06-22 16:27:54 +0300283 tilcdc_crtc_disable(crtc);
Jyri Sarha6c94c712016-09-07 11:46:40 +0300284 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600285
Jyri Sarha4e910c72016-09-06 22:55:33 +0300286 flush_workqueue(priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600287
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300288 of_node_put(crtc->port);
Rob Clark16ea9752013-01-08 15:04:28 -0600289 drm_crtc_cleanup(crtc);
Rob Clarka464d612013-08-07 13:41:20 -0400290 drm_flip_work_cleanup(&tilcdc_crtc->unref_work);
Rob Clark16ea9752013-01-08 15:04:28 -0600291}
292
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300293int tilcdc_crtc_update_fb(struct drm_crtc *crtc,
Rob Clark16ea9752013-01-08 15:04:28 -0600294 struct drm_framebuffer *fb,
Jyri Sarhae0e344e2016-06-22 17:21:06 +0300295 struct drm_pending_vblank_event *event)
Rob Clark16ea9752013-01-08 15:04:28 -0600296{
297 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
298 struct drm_device *dev = crtc->dev;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300299 unsigned long flags;
Tomi Valkeinen6f206e92014-02-07 17:37:07 +0000300
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300301 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
302
Rob Clark16ea9752013-01-08 15:04:28 -0600303 if (tilcdc_crtc->event) {
304 dev_err(dev->dev, "already pending page flip!\n");
305 return -EBUSY;
306 }
307
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300308 drm_framebuffer_reference(fb);
309
Matt Roperf4510a22014-04-01 15:22:40 -0700310 crtc->primary->fb = fb;
Tomi Valkeinen65734a22015-10-19 12:30:03 +0300311
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200312 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300313
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300314 if (crtc->hwmode.vrefresh && ktime_to_ns(tilcdc_crtc->last_vblank)) {
315 ktime_t next_vblank;
316 s64 tdiff;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300317
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300318 next_vblank = ktime_add_us(tilcdc_crtc->last_vblank,
319 1000000 / crtc->hwmode.vrefresh);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200320
Jyri Sarha0a1fe1b2016-06-13 09:53:36 +0300321 tdiff = ktime_to_us(ktime_sub(next_vblank, ktime_get()));
322
323 if (tdiff < TILCDC_VBLANK_SAFETY_THRESHOLD_US)
324 tilcdc_crtc->next_fb = fb;
325 }
326
327 if (tilcdc_crtc->next_fb != fb)
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200328 set_scanout(crtc, fb);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200329
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300330 tilcdc_crtc->event = event;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200331
332 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600333
334 return 0;
335}
336
Rob Clark16ea9752013-01-08 15:04:28 -0600337static bool tilcdc_crtc_mode_fixup(struct drm_crtc *crtc,
338 const struct drm_display_mode *mode,
339 struct drm_display_mode *adjusted_mode)
340{
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200341 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
342
343 if (!tilcdc_crtc->simulate_vesa_sync)
344 return true;
345
346 /*
347 * tilcdc does not generate VESA-compliant sync but aligns
348 * VS on the second edge of HS instead of first edge.
349 * We use adjusted_mode, to fixup sync by aligning both rising
350 * edges and add HSKEW offset to fix the sync.
351 */
352 adjusted_mode->hskew = mode->hsync_end - mode->hsync_start;
353 adjusted_mode->flags |= DRM_MODE_FLAG_HSKEW;
354
355 if (mode->flags & DRM_MODE_FLAG_NHSYNC) {
356 adjusted_mode->flags |= DRM_MODE_FLAG_PHSYNC;
357 adjusted_mode->flags &= ~DRM_MODE_FLAG_NHSYNC;
358 } else {
359 adjusted_mode->flags |= DRM_MODE_FLAG_NHSYNC;
360 adjusted_mode->flags &= ~DRM_MODE_FLAG_PHSYNC;
361 }
362
Rob Clark16ea9752013-01-08 15:04:28 -0600363 return true;
364}
365
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200366/*
367 * Calculate the percentage difference between the requested pixel clock rate
368 * and the effective rate resulting from calculating the clock divider value.
369 */
370static unsigned int tilcdc_pclk_diff(unsigned long rate,
371 unsigned long real_rate)
372{
373 int r = rate / 100, rr = real_rate / 100;
374
375 return (unsigned int)(abs(((rr - r) * 100) / r));
376}
377
Jyri Sarha642e5162016-09-06 16:19:54 +0300378static void tilcdc_crtc_set_clk(struct drm_crtc *crtc)
379{
380 struct drm_device *dev = crtc->dev;
381 struct tilcdc_drm_private *priv = dev->dev_private;
382 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200383 unsigned long clk_rate, real_rate, req_rate;
384 unsigned int clkdiv;
Jyri Sarha642e5162016-09-06 16:19:54 +0300385 int ret;
386
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200387 clkdiv = 2; /* first try using a standard divider of 2 */
388
Jyri Sarha642e5162016-09-06 16:19:54 +0300389 /* mode.clock is in KHz, set_rate wants parameter in Hz */
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200390 req_rate = crtc->mode.clock * 1000;
391
392 ret = clk_set_rate(priv->clk, req_rate * clkdiv);
393 clk_rate = clk_get_rate(priv->clk);
Jyri Sarha642e5162016-09-06 16:19:54 +0300394 if (ret < 0) {
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200395 /*
396 * If we fail to set the clock rate (some architectures don't
397 * use the common clock framework yet and may not implement
398 * all the clk API calls for every clock), try the next best
399 * thing: adjusting the clock divider, unless clk_get_rate()
400 * failed as well.
401 */
402 if (!clk_rate) {
403 /* Nothing more we can do. Just bail out. */
404 dev_err(dev->dev,
405 "failed to set the pixel clock - unable to read current lcdc clock rate\n");
406 return;
407 }
408
409 clkdiv = DIV_ROUND_CLOSEST(clk_rate, req_rate);
410
411 /*
412 * Emit a warning if the real clock rate resulting from the
413 * calculated divider differs much from the requested rate.
414 *
415 * 5% is an arbitrary value - LCDs are usually quite tolerant
416 * about pixel clock rates.
417 */
418 real_rate = clkdiv * req_rate;
419
420 if (tilcdc_pclk_diff(clk_rate, real_rate) > 5) {
421 dev_warn(dev->dev,
422 "effective pixel clock rate (%luHz) differs from the calculated rate (%luHz)\n",
423 clk_rate, real_rate);
424 }
Jyri Sarha642e5162016-09-06 16:19:54 +0300425 }
426
Bartosz Golaszewskicb42e202016-09-29 18:43:57 +0200427 tilcdc_crtc->lcd_fck_rate = clk_rate;
Jyri Sarha642e5162016-09-06 16:19:54 +0300428
429 DBG("lcd_clk=%u, mode clock=%d, div=%u",
430 tilcdc_crtc->lcd_fck_rate, crtc->mode.clock, clkdiv);
431
432 /* Configure the LCD clock divisor. */
433 tilcdc_write(dev, LCDC_CTRL_REG, LCDC_CLK_DIVISOR(clkdiv) |
434 LCDC_RASTER_MODE);
435
436 if (priv->rev == 2)
437 tilcdc_set(dev, LCDC_CLK_ENABLE_REG,
438 LCDC_V2_DMA_CLK_EN | LCDC_V2_LIDD_CLK_EN |
439 LCDC_V2_CORE_CLK_EN);
440}
441
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300442static void tilcdc_crtc_mode_set_nofb(struct drm_crtc *crtc)
443{
444 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
445 struct drm_device *dev = crtc->dev;
446 struct tilcdc_drm_private *priv = dev->dev_private;
447 const struct tilcdc_panel_info *info = tilcdc_crtc->info;
448 uint32_t reg, hbp, hfp, hsw, vbp, vfp, vsw;
449 struct drm_display_mode *mode = &crtc->state->adjusted_mode;
450 struct drm_framebuffer *fb = crtc->primary->state->fb;
451
Jyri Sarha2e0965b2016-09-06 17:25:08 +0300452 WARN_ON(!drm_modeset_is_locked(&crtc->mutex));
453
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300454 if (WARN_ON(!info))
455 return;
456
457 if (WARN_ON(!fb))
458 return;
459
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300460 /* Configure the Burst Size and fifo threshold of DMA: */
461 reg = tilcdc_read(dev, LCDC_DMA_CTRL_REG) & ~0x00000770;
462 switch (info->dma_burst_sz) {
463 case 1:
464 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_1);
465 break;
466 case 2:
467 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_2);
468 break;
469 case 4:
470 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_4);
471 break;
472 case 8:
473 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_8);
474 break;
475 case 16:
476 reg |= LCDC_DMA_BURST_SIZE(LCDC_DMA_BURST_16);
477 break;
478 default:
479 dev_err(dev->dev, "invalid burst size\n");
480 return;
481 }
482 reg |= (info->fifo_th << 8);
483 tilcdc_write(dev, LCDC_DMA_CTRL_REG, reg);
484
485 /* Configure timings: */
486 hbp = mode->htotal - mode->hsync_end;
487 hfp = mode->hsync_start - mode->hdisplay;
488 hsw = mode->hsync_end - mode->hsync_start;
489 vbp = mode->vtotal - mode->vsync_end;
490 vfp = mode->vsync_start - mode->vdisplay;
491 vsw = mode->vsync_end - mode->vsync_start;
492
493 DBG("%dx%d, hbp=%u, hfp=%u, hsw=%u, vbp=%u, vfp=%u, vsw=%u",
494 mode->hdisplay, mode->vdisplay, hbp, hfp, hsw, vbp, vfp, vsw);
495
496 /* Set AC Bias Period and Number of Transitions per Interrupt: */
497 reg = tilcdc_read(dev, LCDC_RASTER_TIMING_2_REG) & ~0x000fff00;
498 reg |= LCDC_AC_BIAS_FREQUENCY(info->ac_bias) |
499 LCDC_AC_BIAS_TRANSITIONS_PER_INT(info->ac_bias_intrpt);
500
501 /*
502 * subtract one from hfp, hbp, hsw because the hardware uses
503 * a value of 0 as 1
504 */
505 if (priv->rev == 2) {
506 /* clear bits we're going to set */
507 reg &= ~0x78000033;
508 reg |= ((hfp-1) & 0x300) >> 8;
509 reg |= ((hbp-1) & 0x300) >> 4;
510 reg |= ((hsw-1) & 0x3c0) << 21;
511 }
512 tilcdc_write(dev, LCDC_RASTER_TIMING_2_REG, reg);
513
514 reg = (((mode->hdisplay >> 4) - 1) << 4) |
515 (((hbp-1) & 0xff) << 24) |
516 (((hfp-1) & 0xff) << 16) |
517 (((hsw-1) & 0x3f) << 10);
518 if (priv->rev == 2)
519 reg |= (((mode->hdisplay >> 4) - 1) & 0x40) >> 3;
520 tilcdc_write(dev, LCDC_RASTER_TIMING_0_REG, reg);
521
522 reg = ((mode->vdisplay - 1) & 0x3ff) |
523 ((vbp & 0xff) << 24) |
524 ((vfp & 0xff) << 16) |
525 (((vsw-1) & 0x3f) << 10);
526 tilcdc_write(dev, LCDC_RASTER_TIMING_1_REG, reg);
527
528 /*
529 * be sure to set Bit 10 for the V2 LCDC controller,
530 * otherwise limited to 1024 pixels width, stopping
531 * 1920x1080 being supported.
532 */
533 if (priv->rev == 2) {
534 if ((mode->vdisplay - 1) & 0x400) {
535 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG,
536 LCDC_LPP_B10);
537 } else {
538 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG,
539 LCDC_LPP_B10);
540 }
541 }
542
543 /* Configure display type: */
544 reg = tilcdc_read(dev, LCDC_RASTER_CTRL_REG) &
545 ~(LCDC_TFT_MODE | LCDC_MONO_8BIT_MODE | LCDC_MONOCHROME_MODE |
546 LCDC_V2_TFT_24BPP_MODE | LCDC_V2_TFT_24BPP_UNPACK |
547 0x000ff000 /* Palette Loading Delay bits */);
548 reg |= LCDC_TFT_MODE; /* no monochrome/passive support */
549 if (info->tft_alt_mode)
550 reg |= LCDC_TFT_ALT_ENABLE;
551 if (priv->rev == 2) {
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300552 switch (fb->pixel_format) {
553 case DRM_FORMAT_BGR565:
554 case DRM_FORMAT_RGB565:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300555 break;
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300556 case DRM_FORMAT_XBGR8888:
557 case DRM_FORMAT_XRGB8888:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300558 reg |= LCDC_V2_TFT_24BPP_UNPACK;
559 /* fallthrough */
Laurent Pinchart59f11a42016-10-18 01:41:14 +0300560 case DRM_FORMAT_BGR888:
561 case DRM_FORMAT_RGB888:
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300562 reg |= LCDC_V2_TFT_24BPP_MODE;
563 break;
564 default:
565 dev_err(dev->dev, "invalid pixel format\n");
566 return;
567 }
568 }
569 reg |= info->fdd < 12;
570 tilcdc_write(dev, LCDC_RASTER_CTRL_REG, reg);
571
572 if (info->invert_pxl_clk)
573 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
574 else
575 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_PIXEL_CLOCK);
576
577 if (info->sync_ctrl)
578 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
579 else
580 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_CTRL);
581
582 if (info->sync_edge)
583 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
584 else
585 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_SYNC_EDGE);
586
587 if (mode->flags & DRM_MODE_FLAG_NHSYNC)
588 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
589 else
590 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_HSYNC);
591
592 if (mode->flags & DRM_MODE_FLAG_NVSYNC)
593 tilcdc_set(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
594 else
595 tilcdc_clear(dev, LCDC_RASTER_TIMING_2_REG, LCDC_INVERT_VSYNC);
596
597 if (info->raster_order)
598 tilcdc_set(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
599 else
600 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG, LCDC_RASTER_ORDER);
601
602 drm_framebuffer_reference(fb);
603
604 set_scanout(crtc, fb);
605
Jyri Sarha642e5162016-09-06 16:19:54 +0300606 tilcdc_crtc_set_clk(crtc);
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300607
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300608 crtc->hwmode = crtc->state->adjusted_mode;
609}
610
Jyri Sarhadb380c52016-04-07 15:10:23 +0300611static int tilcdc_crtc_atomic_check(struct drm_crtc *crtc,
612 struct drm_crtc_state *state)
613{
614 struct drm_display_mode *mode = &state->mode;
615 int ret;
616
617 /* If we are not active we don't care */
618 if (!state->active)
619 return 0;
620
621 if (state->state->planes[0].ptr != crtc->primary ||
622 state->state->planes[0].state == NULL ||
623 state->state->planes[0].state->crtc != crtc) {
624 dev_dbg(crtc->dev->dev, "CRTC primary plane must be present");
625 return -EINVAL;
626 }
627
628 ret = tilcdc_crtc_mode_valid(crtc, mode);
629 if (ret) {
630 dev_dbg(crtc->dev->dev, "Mode \"%s\" not valid", mode->name);
631 return -EINVAL;
632 }
633
634 return 0;
635}
636
Rob Clark16ea9752013-01-08 15:04:28 -0600637static const struct drm_crtc_funcs tilcdc_crtc_funcs = {
Jyri Sarha305198d2016-04-07 15:05:16 +0300638 .destroy = tilcdc_crtc_destroy,
639 .set_config = drm_atomic_helper_set_config,
640 .page_flip = drm_atomic_helper_page_flip,
641 .reset = drm_atomic_helper_crtc_reset,
642 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
643 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
Rob Clark16ea9752013-01-08 15:04:28 -0600644};
645
646static const struct drm_crtc_helper_funcs tilcdc_crtc_helper_funcs = {
Rob Clark16ea9752013-01-08 15:04:28 -0600647 .mode_fixup = tilcdc_crtc_mode_fixup,
Jyri Sarha305198d2016-04-07 15:05:16 +0300648 .enable = tilcdc_crtc_enable,
649 .disable = tilcdc_crtc_disable,
Jyri Sarhadb380c52016-04-07 15:10:23 +0300650 .atomic_check = tilcdc_crtc_atomic_check,
Jyri Sarhaf6382f12016-04-07 15:09:50 +0300651 .mode_set_nofb = tilcdc_crtc_mode_set_nofb,
Rob Clark16ea9752013-01-08 15:04:28 -0600652};
653
654int tilcdc_crtc_max_width(struct drm_crtc *crtc)
655{
656 struct drm_device *dev = crtc->dev;
657 struct tilcdc_drm_private *priv = dev->dev_private;
658 int max_width = 0;
659
660 if (priv->rev == 1)
661 max_width = 1024;
662 else if (priv->rev == 2)
663 max_width = 2048;
664
665 return max_width;
666}
667
668int tilcdc_crtc_mode_valid(struct drm_crtc *crtc, struct drm_display_mode *mode)
669{
670 struct tilcdc_drm_private *priv = crtc->dev->dev_private;
671 unsigned int bandwidth;
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500672 uint32_t hbp, hfp, hsw, vbp, vfp, vsw;
Rob Clark16ea9752013-01-08 15:04:28 -0600673
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500674 /*
675 * check to see if the width is within the range that
676 * the LCD Controller physically supports
677 */
Rob Clark16ea9752013-01-08 15:04:28 -0600678 if (mode->hdisplay > tilcdc_crtc_max_width(crtc))
679 return MODE_VIRTUAL_X;
680
681 /* width must be multiple of 16 */
682 if (mode->hdisplay & 0xf)
683 return MODE_VIRTUAL_X;
684
685 if (mode->vdisplay > 2048)
686 return MODE_VIRTUAL_Y;
687
Darren Etheridgee1c5d0a2013-06-21 13:52:25 -0500688 DBG("Processing mode %dx%d@%d with pixel clock %d",
689 mode->hdisplay, mode->vdisplay,
690 drm_mode_vrefresh(mode), mode->clock);
691
692 hbp = mode->htotal - mode->hsync_end;
693 hfp = mode->hsync_start - mode->hdisplay;
694 hsw = mode->hsync_end - mode->hsync_start;
695 vbp = mode->vtotal - mode->vsync_end;
696 vfp = mode->vsync_start - mode->vdisplay;
697 vsw = mode->vsync_end - mode->vsync_start;
698
699 if ((hbp-1) & ~0x3ff) {
700 DBG("Pruning mode: Horizontal Back Porch out of range");
701 return MODE_HBLANK_WIDE;
702 }
703
704 if ((hfp-1) & ~0x3ff) {
705 DBG("Pruning mode: Horizontal Front Porch out of range");
706 return MODE_HBLANK_WIDE;
707 }
708
709 if ((hsw-1) & ~0x3ff) {
710 DBG("Pruning mode: Horizontal Sync Width out of range");
711 return MODE_HSYNC_WIDE;
712 }
713
714 if (vbp & ~0xff) {
715 DBG("Pruning mode: Vertical Back Porch out of range");
716 return MODE_VBLANK_WIDE;
717 }
718
719 if (vfp & ~0xff) {
720 DBG("Pruning mode: Vertical Front Porch out of range");
721 return MODE_VBLANK_WIDE;
722 }
723
724 if ((vsw-1) & ~0x3f) {
725 DBG("Pruning mode: Vertical Sync Width out of range");
726 return MODE_VSYNC_WIDE;
727 }
728
Darren Etheridge4e564342013-06-21 13:52:23 -0500729 /*
730 * some devices have a maximum allowed pixel clock
731 * configured from the DT
732 */
733 if (mode->clock > priv->max_pixelclock) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500734 DBG("Pruning mode: pixel clock too high");
Darren Etheridge4e564342013-06-21 13:52:23 -0500735 return MODE_CLOCK_HIGH;
736 }
737
738 /*
739 * some devices further limit the max horizontal resolution
740 * configured from the DT
741 */
742 if (mode->hdisplay > priv->max_width)
743 return MODE_BAD_WIDTH;
744
Rob Clark16ea9752013-01-08 15:04:28 -0600745 /* filter out modes that would require too much memory bandwidth: */
Darren Etheridge4e564342013-06-21 13:52:23 -0500746 bandwidth = mode->hdisplay * mode->vdisplay *
747 drm_mode_vrefresh(mode);
748 if (bandwidth > priv->max_bandwidth) {
Darren Etheridgef7b45752013-06-21 13:52:26 -0500749 DBG("Pruning mode: exceeds defined bandwidth limit");
Rob Clark16ea9752013-01-08 15:04:28 -0600750 return MODE_BAD;
Darren Etheridge4e564342013-06-21 13:52:23 -0500751 }
Rob Clark16ea9752013-01-08 15:04:28 -0600752
753 return MODE_OK;
754}
755
756void tilcdc_crtc_set_panel_info(struct drm_crtc *crtc,
757 const struct tilcdc_panel_info *info)
758{
759 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
760 tilcdc_crtc->info = info;
761}
762
Jyri Sarha103cd8b2015-02-10 14:13:23 +0200763void tilcdc_crtc_set_simulate_vesa_sync(struct drm_crtc *crtc,
764 bool simulate_vesa_sync)
765{
766 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
767
768 tilcdc_crtc->simulate_vesa_sync = simulate_vesa_sync;
769}
770
Rob Clark16ea9752013-01-08 15:04:28 -0600771void tilcdc_crtc_update_clk(struct drm_crtc *crtc)
772{
Rob Clark16ea9752013-01-08 15:04:28 -0600773 struct drm_device *dev = crtc->dev;
774 struct tilcdc_drm_private *priv = dev->dev_private;
Jyri Sarha642e5162016-09-06 16:19:54 +0300775 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600776
Jyri Sarha642e5162016-09-06 16:19:54 +0300777 drm_modeset_lock_crtc(crtc, NULL);
778 if (tilcdc_crtc->lcd_fck_rate != clk_get_rate(priv->clk)) {
779 if (tilcdc_crtc_is_on(crtc)) {
780 pm_runtime_get_sync(dev->dev);
781 tilcdc_crtc_disable(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600782
Jyri Sarha642e5162016-09-06 16:19:54 +0300783 tilcdc_crtc_set_clk(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600784
Jyri Sarha642e5162016-09-06 16:19:54 +0300785 tilcdc_crtc_enable(crtc);
786 pm_runtime_put_sync(dev->dev);
787 }
Rob Clark16ea9752013-01-08 15:04:28 -0600788 }
Jyri Sarha642e5162016-09-06 16:19:54 +0300789 drm_modeset_unlock_crtc(crtc);
Rob Clark16ea9752013-01-08 15:04:28 -0600790}
791
Jyri Sarha5895d082016-01-08 14:33:09 +0200792#define SYNC_LOST_COUNT_LIMIT 50
793
Rob Clark16ea9752013-01-08 15:04:28 -0600794irqreturn_t tilcdc_crtc_irq(struct drm_crtc *crtc)
795{
796 struct tilcdc_crtc *tilcdc_crtc = to_tilcdc_crtc(crtc);
797 struct drm_device *dev = crtc->dev;
798 struct tilcdc_drm_private *priv = dev->dev_private;
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300799 uint32_t stat;
Rob Clark16ea9752013-01-08 15:04:28 -0600800
Tomi Valkeinen317aae72015-10-20 12:08:03 +0300801 stat = tilcdc_read_irqstatus(dev);
802 tilcdc_clear_irqstatus(dev, stat);
803
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300804 if (stat & LCDC_END_OF_FRAME0) {
Rob Clark16ea9752013-01-08 15:04:28 -0600805 unsigned long flags;
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200806 bool skip_event = false;
807 ktime_t now;
808
809 now = ktime_get();
Rob Clark16ea9752013-01-08 15:04:28 -0600810
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300811 drm_flip_work_commit(&tilcdc_crtc->unref_work, priv->wq);
Rob Clark16ea9752013-01-08 15:04:28 -0600812
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200813 spin_lock_irqsave(&tilcdc_crtc->irq_lock, flags);
Rob Clark16ea9752013-01-08 15:04:28 -0600814
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200815 tilcdc_crtc->last_vblank = now;
Rob Clark16ea9752013-01-08 15:04:28 -0600816
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200817 if (tilcdc_crtc->next_fb) {
818 set_scanout(crtc, tilcdc_crtc->next_fb);
819 tilcdc_crtc->next_fb = NULL;
820 skip_event = true;
Tomi Valkeinen2b2080d72015-10-20 09:37:27 +0300821 }
822
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200823 spin_unlock_irqrestore(&tilcdc_crtc->irq_lock, flags);
824
Gustavo Padovan099ede82016-07-04 21:04:52 -0300825 drm_crtc_handle_vblank(crtc);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200826
827 if (!skip_event) {
828 struct drm_pending_vblank_event *event;
829
830 spin_lock_irqsave(&dev->event_lock, flags);
831
832 event = tilcdc_crtc->event;
833 tilcdc_crtc->event = NULL;
834 if (event)
Gustavo Padovandfebc152016-04-14 10:48:22 -0700835 drm_crtc_send_vblank_event(crtc, event);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200836
837 spin_unlock_irqrestore(&dev->event_lock, flags);
838 }
Jyri Sarha5895d082016-01-08 14:33:09 +0200839
840 if (tilcdc_crtc->frame_intact)
841 tilcdc_crtc->sync_lost_count = 0;
842 else
843 tilcdc_crtc->frame_intact = true;
Rob Clark16ea9752013-01-08 15:04:28 -0600844 }
845
Jyri Sarha14944112016-04-07 20:36:48 +0300846 if (stat & LCDC_FIFO_UNDERFLOW)
Daniel Schultzd7014532016-10-28 13:52:42 +0200847 dev_err_ratelimited(dev->dev, "%s(0x%08x): FIFO underflow",
Jyri Sarha14944112016-04-07 20:36:48 +0300848 __func__, stat);
849
Jyri Sarhacba88442016-11-16 00:12:27 +0200850 if (stat & LCDC_SYNC_LOST) {
851 dev_err_ratelimited(dev->dev, "%s(0x%08x): Sync lost",
852 __func__, stat);
853 tilcdc_crtc->frame_intact = false;
854 if (tilcdc_crtc->sync_lost_count++ >
855 SYNC_LOST_COUNT_LIMIT) {
856 dev_err(dev->dev, "%s(0x%08x): Sync lost flood detected, recovering", __func__, stat);
857 queue_work(system_wq, &tilcdc_crtc->recover_work);
858 if (priv->rev == 1)
859 tilcdc_clear(dev, LCDC_RASTER_CTRL_REG,
860 LCDC_V1_SYNC_LOST_INT_ENA);
861 else
862 tilcdc_write(dev, LCDC_INT_ENABLE_CLR_REG,
863 LCDC_SYNC_LOST);
864 tilcdc_crtc->sync_lost_count = 0;
865 }
866 }
867
Jyri Sarha14944112016-04-07 20:36:48 +0300868 /* For revision 2 only */
Rob Clark16ea9752013-01-08 15:04:28 -0600869 if (priv->rev == 2) {
870 if (stat & LCDC_FRAME_DONE) {
871 tilcdc_crtc->frame_done = true;
872 wake_up(&tilcdc_crtc->frame_done_wq);
873 }
Rob Clark16ea9752013-01-08 15:04:28 -0600874
Jyri Sarha14944112016-04-07 20:36:48 +0300875 /* Indicate to LCDC that the interrupt service routine has
876 * completed, see 13.3.6.1.6 in AM335x TRM.
877 */
878 tilcdc_write(dev, LCDC_END_OF_INT_IND_REG, 0);
879 }
Jyri Sarhac0c2baa2015-12-18 13:07:52 +0200880
Rob Clark16ea9752013-01-08 15:04:28 -0600881 return IRQ_HANDLED;
882}
883
Rob Clark16ea9752013-01-08 15:04:28 -0600884struct drm_crtc *tilcdc_crtc_create(struct drm_device *dev)
885{
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300886 struct tilcdc_drm_private *priv = dev->dev_private;
Rob Clark16ea9752013-01-08 15:04:28 -0600887 struct tilcdc_crtc *tilcdc_crtc;
888 struct drm_crtc *crtc;
889 int ret;
890
Jyri Sarhad0ec32c2016-02-23 12:44:27 +0200891 tilcdc_crtc = devm_kzalloc(dev->dev, sizeof(*tilcdc_crtc), GFP_KERNEL);
Rob Clark16ea9752013-01-08 15:04:28 -0600892 if (!tilcdc_crtc) {
893 dev_err(dev->dev, "allocation failed\n");
894 return NULL;
895 }
896
897 crtc = &tilcdc_crtc->base;
898
Jyri Sarha47f571c2016-04-07 15:04:18 +0300899 ret = tilcdc_plane_init(dev, &tilcdc_crtc->primary);
900 if (ret < 0)
901 goto fail;
902
Jyri Sarha2d53a182016-10-25 12:27:31 +0300903 mutex_init(&tilcdc_crtc->enable_lock);
904
Rob Clark16ea9752013-01-08 15:04:28 -0600905 init_waitqueue_head(&tilcdc_crtc->frame_done_wq);
906
Boris BREZILLONd7f8db52014-11-14 19:30:30 +0100907 drm_flip_work_init(&tilcdc_crtc->unref_work,
Rob Clarka464d612013-08-07 13:41:20 -0400908 "unref", unref_worker);
Rob Clark16ea9752013-01-08 15:04:28 -0600909
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200910 spin_lock_init(&tilcdc_crtc->irq_lock);
Jyri Sarha13b3d722016-04-06 14:02:38 +0300911 INIT_WORK(&tilcdc_crtc->recover_work, tilcdc_crtc_recover_work);
Tomi Valkeinen2b3a8cd2015-11-03 12:00:51 +0200912
Jyri Sarha47f571c2016-04-07 15:04:18 +0300913 ret = drm_crtc_init_with_planes(dev, crtc,
914 &tilcdc_crtc->primary,
915 NULL,
916 &tilcdc_crtc_funcs,
917 "tilcdc crtc");
Rob Clark16ea9752013-01-08 15:04:28 -0600918 if (ret < 0)
919 goto fail;
920
921 drm_crtc_helper_add(crtc, &tilcdc_crtc_helper_funcs);
922
Jyri Sarhad66284fb2015-05-27 11:58:37 +0300923 if (priv->is_componentized) {
924 struct device_node *ports =
925 of_get_child_by_name(dev->dev->of_node, "ports");
926
927 if (ports) {
928 crtc->port = of_get_child_by_name(ports, "port");
929 of_node_put(ports);
930 } else {
931 crtc->port =
932 of_get_child_by_name(dev->dev->of_node, "port");
933 }
934 if (!crtc->port) { /* This should never happen */
935 dev_err(dev->dev, "Port node not found in %s\n",
936 dev->dev->of_node->full_name);
937 goto fail;
938 }
939 }
940
Rob Clark16ea9752013-01-08 15:04:28 -0600941 return crtc;
942
943fail:
944 tilcdc_crtc_destroy(crtc);
945 return NULL;
946}