| /* |
| * linux/drivers/video/omap2/dss/dss.c |
| * |
| * Copyright (C) 2009 Nokia Corporation |
| * Author: Tomi Valkeinen <tomi.valkeinen@nokia.com> |
| * |
| * Some code and ideas taken from drivers/video/omap/ driver |
| * by Imre Deak. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License version 2 as published by |
| * the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #define DSS_SUBSYS_NAME "DSS" |
| |
| #include <linux/kernel.h> |
| #include <linux/io.h> |
| #include <linux/err.h> |
| #include <linux/delay.h> |
| #include <linux/interrupt.h> |
| #include <linux/seq_file.h> |
| #include <linux/clk.h> |
| |
| #include <plat/display.h> |
| #include "dss.h" |
| |
| #define DSS_BASE 0x48050000 |
| |
| #define DSS_SZ_REGS SZ_512 |
| |
| struct dss_reg { |
| u16 idx; |
| }; |
| |
| #define DSS_REG(idx) ((const struct dss_reg) { idx }) |
| |
| #define DSS_REVISION DSS_REG(0x0000) |
| #define DSS_SYSCONFIG DSS_REG(0x0010) |
| #define DSS_SYSSTATUS DSS_REG(0x0014) |
| #define DSS_IRQSTATUS DSS_REG(0x0018) |
| #define DSS_CONTROL DSS_REG(0x0040) |
| #define DSS_SDI_CONTROL DSS_REG(0x0044) |
| #define DSS_PLL_CONTROL DSS_REG(0x0048) |
| #define DSS_SDI_STATUS DSS_REG(0x005C) |
| |
| #define REG_GET(idx, start, end) \ |
| FLD_GET(dss_read_reg(idx), start, end) |
| |
| #define REG_FLD_MOD(idx, val, start, end) \ |
| dss_write_reg(idx, FLD_MOD(dss_read_reg(idx), val, start, end)) |
| |
| static struct { |
| void __iomem *base; |
| |
| struct clk *dpll4_m4_ck; |
| |
| unsigned long cache_req_pck; |
| unsigned long cache_prate; |
| struct dss_clock_info cache_dss_cinfo; |
| struct dispc_clock_info cache_dispc_cinfo; |
| |
| enum dss_clk_source dsi_clk_source; |
| enum dss_clk_source dispc_clk_source; |
| |
| u32 ctx[DSS_SZ_REGS / sizeof(u32)]; |
| } dss; |
| |
| static int _omap_dss_wait_reset(void); |
| |
| static inline void dss_write_reg(const struct dss_reg idx, u32 val) |
| { |
| __raw_writel(val, dss.base + idx.idx); |
| } |
| |
| static inline u32 dss_read_reg(const struct dss_reg idx) |
| { |
| return __raw_readl(dss.base + idx.idx); |
| } |
| |
| #define SR(reg) \ |
| dss.ctx[(DSS_##reg).idx / sizeof(u32)] = dss_read_reg(DSS_##reg) |
| #define RR(reg) \ |
| dss_write_reg(DSS_##reg, dss.ctx[(DSS_##reg).idx / sizeof(u32)]) |
| |
| void dss_save_context(void) |
| { |
| if (cpu_is_omap24xx()) |
| return; |
| |
| SR(SYSCONFIG); |
| SR(CONTROL); |
| |
| #ifdef CONFIG_OMAP2_DSS_SDI |
| SR(SDI_CONTROL); |
| SR(PLL_CONTROL); |
| #endif |
| } |
| |
| void dss_restore_context(void) |
| { |
| if (_omap_dss_wait_reset()) |
| DSSERR("DSS not coming out of reset after sleep\n"); |
| |
| RR(SYSCONFIG); |
| RR(CONTROL); |
| |
| #ifdef CONFIG_OMAP2_DSS_SDI |
| RR(SDI_CONTROL); |
| RR(PLL_CONTROL); |
| #endif |
| } |
| |
| #undef SR |
| #undef RR |
| |
| void dss_sdi_init(u8 datapairs) |
| { |
| u32 l; |
| |
| BUG_ON(datapairs > 3 || datapairs < 1); |
| |
| l = dss_read_reg(DSS_SDI_CONTROL); |
| l = FLD_MOD(l, 0xf, 19, 15); /* SDI_PDIV */ |
| l = FLD_MOD(l, datapairs-1, 3, 2); /* SDI_PRSEL */ |
| l = FLD_MOD(l, 2, 1, 0); /* SDI_BWSEL */ |
| dss_write_reg(DSS_SDI_CONTROL, l); |
| |
| l = dss_read_reg(DSS_PLL_CONTROL); |
| l = FLD_MOD(l, 0x7, 25, 22); /* SDI_PLL_FREQSEL */ |
| l = FLD_MOD(l, 0xb, 16, 11); /* SDI_PLL_REGN */ |
| l = FLD_MOD(l, 0xb4, 10, 1); /* SDI_PLL_REGM */ |
| dss_write_reg(DSS_PLL_CONTROL, l); |
| } |
| |
| int dss_sdi_enable(void) |
| { |
| unsigned long timeout; |
| |
| dispc_pck_free_enable(1); |
| |
| /* Reset SDI PLL */ |
| REG_FLD_MOD(DSS_PLL_CONTROL, 1, 18, 18); /* SDI_PLL_SYSRESET */ |
| udelay(1); /* wait 2x PCLK */ |
| |
| /* Lock SDI PLL */ |
| REG_FLD_MOD(DSS_PLL_CONTROL, 1, 28, 28); /* SDI_PLL_GOBIT */ |
| |
| /* Waiting for PLL lock request to complete */ |
| timeout = jiffies + msecs_to_jiffies(500); |
| while (dss_read_reg(DSS_SDI_STATUS) & (1 << 6)) { |
| if (time_after_eq(jiffies, timeout)) { |
| DSSERR("PLL lock request timed out\n"); |
| goto err1; |
| } |
| } |
| |
| /* Clearing PLL_GO bit */ |
| REG_FLD_MOD(DSS_PLL_CONTROL, 0, 28, 28); |
| |
| /* Waiting for PLL to lock */ |
| timeout = jiffies + msecs_to_jiffies(500); |
| while (!(dss_read_reg(DSS_SDI_STATUS) & (1 << 5))) { |
| if (time_after_eq(jiffies, timeout)) { |
| DSSERR("PLL lock timed out\n"); |
| goto err1; |
| } |
| } |
| |
| dispc_lcd_enable_signal(1); |
| |
| /* Waiting for SDI reset to complete */ |
| timeout = jiffies + msecs_to_jiffies(500); |
| while (!(dss_read_reg(DSS_SDI_STATUS) & (1 << 2))) { |
| if (time_after_eq(jiffies, timeout)) { |
| DSSERR("SDI reset timed out\n"); |
| goto err2; |
| } |
| } |
| |
| return 0; |
| |
| err2: |
| dispc_lcd_enable_signal(0); |
| err1: |
| /* Reset SDI PLL */ |
| REG_FLD_MOD(DSS_PLL_CONTROL, 0, 18, 18); /* SDI_PLL_SYSRESET */ |
| |
| dispc_pck_free_enable(0); |
| |
| return -ETIMEDOUT; |
| } |
| |
| void dss_sdi_disable(void) |
| { |
| dispc_lcd_enable_signal(0); |
| |
| dispc_pck_free_enable(0); |
| |
| /* Reset SDI PLL */ |
| REG_FLD_MOD(DSS_PLL_CONTROL, 0, 18, 18); /* SDI_PLL_SYSRESET */ |
| } |
| |
| void dss_dump_clocks(struct seq_file *s) |
| { |
| unsigned long dpll4_ck_rate; |
| unsigned long dpll4_m4_ck_rate; |
| |
| dss_clk_enable(DSS_CLK_ICK | DSS_CLK_FCK1); |
| |
| dpll4_ck_rate = clk_get_rate(clk_get_parent(dss.dpll4_m4_ck)); |
| dpll4_m4_ck_rate = clk_get_rate(dss.dpll4_m4_ck); |
| |
| seq_printf(s, "- DSS -\n"); |
| |
| seq_printf(s, "dpll4_ck %lu\n", dpll4_ck_rate); |
| |
| if (cpu_is_omap3630()) |
| seq_printf(s, "dss1_alwon_fclk = %lu / %lu = %lu\n", |
| dpll4_ck_rate, |
| dpll4_ck_rate / dpll4_m4_ck_rate, |
| dss_clk_get_rate(DSS_CLK_FCK1)); |
| else |
| seq_printf(s, "dss1_alwon_fclk = %lu / %lu * 2 = %lu\n", |
| dpll4_ck_rate, |
| dpll4_ck_rate / dpll4_m4_ck_rate, |
| dss_clk_get_rate(DSS_CLK_FCK1)); |
| |
| dss_clk_disable(DSS_CLK_ICK | DSS_CLK_FCK1); |
| } |
| |
| void dss_dump_regs(struct seq_file *s) |
| { |
| #define DUMPREG(r) seq_printf(s, "%-35s %08x\n", #r, dss_read_reg(r)) |
| |
| dss_clk_enable(DSS_CLK_ICK | DSS_CLK_FCK1); |
| |
| DUMPREG(DSS_REVISION); |
| DUMPREG(DSS_SYSCONFIG); |
| DUMPREG(DSS_SYSSTATUS); |
| DUMPREG(DSS_IRQSTATUS); |
| DUMPREG(DSS_CONTROL); |
| DUMPREG(DSS_SDI_CONTROL); |
| DUMPREG(DSS_PLL_CONTROL); |
| DUMPREG(DSS_SDI_STATUS); |
| |
| dss_clk_disable(DSS_CLK_ICK | DSS_CLK_FCK1); |
| #undef DUMPREG |
| } |
| |
| void dss_select_dispc_clk_source(enum dss_clk_source clk_src) |
| { |
| int b; |
| |
| BUG_ON(clk_src != DSS_SRC_DSI1_PLL_FCLK && |
| clk_src != DSS_SRC_DSS1_ALWON_FCLK); |
| |
| b = clk_src == DSS_SRC_DSS1_ALWON_FCLK ? 0 : 1; |
| |
| if (clk_src == DSS_SRC_DSI1_PLL_FCLK) |
| dsi_wait_dsi1_pll_active(); |
| |
| REG_FLD_MOD(DSS_CONTROL, b, 0, 0); /* DISPC_CLK_SWITCH */ |
| |
| dss.dispc_clk_source = clk_src; |
| } |
| |
| void dss_select_dsi_clk_source(enum dss_clk_source clk_src) |
| { |
| int b; |
| |
| BUG_ON(clk_src != DSS_SRC_DSI2_PLL_FCLK && |
| clk_src != DSS_SRC_DSS1_ALWON_FCLK); |
| |
| b = clk_src == DSS_SRC_DSS1_ALWON_FCLK ? 0 : 1; |
| |
| if (clk_src == DSS_SRC_DSI2_PLL_FCLK) |
| dsi_wait_dsi2_pll_active(); |
| |
| REG_FLD_MOD(DSS_CONTROL, b, 1, 1); /* DSI_CLK_SWITCH */ |
| |
| dss.dsi_clk_source = clk_src; |
| } |
| |
| enum dss_clk_source dss_get_dispc_clk_source(void) |
| { |
| return dss.dispc_clk_source; |
| } |
| |
| enum dss_clk_source dss_get_dsi_clk_source(void) |
| { |
| return dss.dsi_clk_source; |
| } |
| |
| /* calculate clock rates using dividers in cinfo */ |
| int dss_calc_clock_rates(struct dss_clock_info *cinfo) |
| { |
| unsigned long prate; |
| |
| if (cinfo->fck_div > (cpu_is_omap3630() ? 32 : 16) || |
| cinfo->fck_div == 0) |
| return -EINVAL; |
| |
| prate = clk_get_rate(clk_get_parent(dss.dpll4_m4_ck)); |
| |
| cinfo->fck = prate / cinfo->fck_div; |
| |
| return 0; |
| } |
| |
| int dss_set_clock_div(struct dss_clock_info *cinfo) |
| { |
| unsigned long prate; |
| int r; |
| |
| if (cpu_is_omap34xx()) { |
| prate = clk_get_rate(clk_get_parent(dss.dpll4_m4_ck)); |
| DSSDBG("dpll4_m4 = %ld\n", prate); |
| |
| r = clk_set_rate(dss.dpll4_m4_ck, prate / cinfo->fck_div); |
| if (r) |
| return r; |
| } |
| |
| DSSDBG("fck = %ld (%d)\n", cinfo->fck, cinfo->fck_div); |
| |
| return 0; |
| } |
| |
| int dss_get_clock_div(struct dss_clock_info *cinfo) |
| { |
| cinfo->fck = dss_clk_get_rate(DSS_CLK_FCK1); |
| |
| if (cpu_is_omap34xx()) { |
| unsigned long prate; |
| prate = clk_get_rate(clk_get_parent(dss.dpll4_m4_ck)); |
| if (cpu_is_omap3630()) |
| cinfo->fck_div = prate / (cinfo->fck); |
| else |
| cinfo->fck_div = prate / (cinfo->fck / 2); |
| } else { |
| cinfo->fck_div = 0; |
| } |
| |
| return 0; |
| } |
| |
| unsigned long dss_get_dpll4_rate(void) |
| { |
| if (cpu_is_omap34xx()) |
| return clk_get_rate(clk_get_parent(dss.dpll4_m4_ck)); |
| else |
| return 0; |
| } |
| |
| int dss_calc_clock_div(bool is_tft, unsigned long req_pck, |
| struct dss_clock_info *dss_cinfo, |
| struct dispc_clock_info *dispc_cinfo) |
| { |
| unsigned long prate; |
| struct dss_clock_info best_dss; |
| struct dispc_clock_info best_dispc; |
| |
| unsigned long fck; |
| |
| u16 fck_div; |
| |
| int match = 0; |
| int min_fck_per_pck; |
| |
| prate = dss_get_dpll4_rate(); |
| |
| fck = dss_clk_get_rate(DSS_CLK_FCK1); |
| if (req_pck == dss.cache_req_pck && |
| ((cpu_is_omap34xx() && prate == dss.cache_prate) || |
| dss.cache_dss_cinfo.fck == fck)) { |
| DSSDBG("dispc clock info found from cache.\n"); |
| *dss_cinfo = dss.cache_dss_cinfo; |
| *dispc_cinfo = dss.cache_dispc_cinfo; |
| return 0; |
| } |
| |
| min_fck_per_pck = CONFIG_OMAP2_DSS_MIN_FCK_PER_PCK; |
| |
| if (min_fck_per_pck && |
| req_pck * min_fck_per_pck > DISPC_MAX_FCK) { |
| DSSERR("Requested pixel clock not possible with the current " |
| "OMAP2_DSS_MIN_FCK_PER_PCK setting. Turning " |
| "the constraint off.\n"); |
| min_fck_per_pck = 0; |
| } |
| |
| retry: |
| memset(&best_dss, 0, sizeof(best_dss)); |
| memset(&best_dispc, 0, sizeof(best_dispc)); |
| |
| if (cpu_is_omap24xx()) { |
| struct dispc_clock_info cur_dispc; |
| /* XXX can we change the clock on omap2? */ |
| fck = dss_clk_get_rate(DSS_CLK_FCK1); |
| fck_div = 1; |
| |
| dispc_find_clk_divs(is_tft, req_pck, fck, &cur_dispc); |
| match = 1; |
| |
| best_dss.fck = fck; |
| best_dss.fck_div = fck_div; |
| |
| best_dispc = cur_dispc; |
| |
| goto found; |
| } else if (cpu_is_omap34xx()) { |
| for (fck_div = (cpu_is_omap3630() ? 32 : 16); |
| fck_div > 0; --fck_div) { |
| struct dispc_clock_info cur_dispc; |
| |
| if (cpu_is_omap3630()) |
| fck = prate / fck_div; |
| else |
| fck = prate / fck_div * 2; |
| |
| if (fck > DISPC_MAX_FCK) |
| continue; |
| |
| if (min_fck_per_pck && |
| fck < req_pck * min_fck_per_pck) |
| continue; |
| |
| match = 1; |
| |
| dispc_find_clk_divs(is_tft, req_pck, fck, &cur_dispc); |
| |
| if (abs(cur_dispc.pck - req_pck) < |
| abs(best_dispc.pck - req_pck)) { |
| |
| best_dss.fck = fck; |
| best_dss.fck_div = fck_div; |
| |
| best_dispc = cur_dispc; |
| |
| if (cur_dispc.pck == req_pck) |
| goto found; |
| } |
| } |
| } else { |
| BUG(); |
| } |
| |
| found: |
| if (!match) { |
| if (min_fck_per_pck) { |
| DSSERR("Could not find suitable clock settings.\n" |
| "Turning FCK/PCK constraint off and" |
| "trying again.\n"); |
| min_fck_per_pck = 0; |
| goto retry; |
| } |
| |
| DSSERR("Could not find suitable clock settings.\n"); |
| |
| return -EINVAL; |
| } |
| |
| if (dss_cinfo) |
| *dss_cinfo = best_dss; |
| if (dispc_cinfo) |
| *dispc_cinfo = best_dispc; |
| |
| dss.cache_req_pck = req_pck; |
| dss.cache_prate = prate; |
| dss.cache_dss_cinfo = best_dss; |
| dss.cache_dispc_cinfo = best_dispc; |
| |
| return 0; |
| } |
| |
| |
| |
| static irqreturn_t dss_irq_handler_omap2(int irq, void *arg) |
| { |
| dispc_irq_handler(); |
| |
| return IRQ_HANDLED; |
| } |
| |
| static irqreturn_t dss_irq_handler_omap3(int irq, void *arg) |
| { |
| u32 irqstatus; |
| |
| irqstatus = dss_read_reg(DSS_IRQSTATUS); |
| |
| if (irqstatus & (1<<0)) /* DISPC_IRQ */ |
| dispc_irq_handler(); |
| #ifdef CONFIG_OMAP2_DSS_DSI |
| if (irqstatus & (1<<1)) /* DSI_IRQ */ |
| dsi_irq_handler(); |
| #endif |
| |
| return IRQ_HANDLED; |
| } |
| |
| static int _omap_dss_wait_reset(void) |
| { |
| int t = 0; |
| |
| while (REG_GET(DSS_SYSSTATUS, 0, 0) == 0) { |
| if (++t > 1000) { |
| DSSERR("soft reset failed\n"); |
| return -ENODEV; |
| } |
| udelay(1); |
| } |
| |
| return 0; |
| } |
| |
| static int _omap_dss_reset(void) |
| { |
| /* Soft reset */ |
| REG_FLD_MOD(DSS_SYSCONFIG, 1, 1, 1); |
| return _omap_dss_wait_reset(); |
| } |
| |
| void dss_set_venc_output(enum omap_dss_venc_type type) |
| { |
| int l = 0; |
| |
| if (type == OMAP_DSS_VENC_TYPE_COMPOSITE) |
| l = 0; |
| else if (type == OMAP_DSS_VENC_TYPE_SVIDEO) |
| l = 1; |
| else |
| BUG(); |
| |
| /* venc out selection. 0 = comp, 1 = svideo */ |
| REG_FLD_MOD(DSS_CONTROL, l, 6, 6); |
| } |
| |
| void dss_set_dac_pwrdn_bgz(bool enable) |
| { |
| REG_FLD_MOD(DSS_CONTROL, enable, 5, 5); /* DAC Power-Down Control */ |
| } |
| |
| int dss_init(bool skip_init) |
| { |
| int r; |
| u32 rev; |
| |
| dss.base = ioremap(DSS_BASE, DSS_SZ_REGS); |
| if (!dss.base) { |
| DSSERR("can't ioremap DSS\n"); |
| r = -ENOMEM; |
| goto fail0; |
| } |
| |
| if (!skip_init) { |
| /* disable LCD and DIGIT output. This seems to fix the synclost |
| * problem that we get, if the bootloader starts the DSS and |
| * the kernel resets it */ |
| omap_writel(omap_readl(0x48050440) & ~0x3, 0x48050440); |
| |
| /* We need to wait here a bit, otherwise we sometimes start to |
| * get synclost errors, and after that only power cycle will |
| * restore DSS functionality. I have no idea why this happens. |
| * And we have to wait _before_ resetting the DSS, but after |
| * enabling clocks. |
| */ |
| msleep(50); |
| |
| _omap_dss_reset(); |
| } |
| |
| /* autoidle */ |
| REG_FLD_MOD(DSS_SYSCONFIG, 1, 0, 0); |
| |
| /* Select DPLL */ |
| REG_FLD_MOD(DSS_CONTROL, 0, 0, 0); |
| |
| #ifdef CONFIG_OMAP2_DSS_VENC |
| REG_FLD_MOD(DSS_CONTROL, 1, 4, 4); /* venc dac demen */ |
| REG_FLD_MOD(DSS_CONTROL, 1, 3, 3); /* venc clock 4x enable */ |
| REG_FLD_MOD(DSS_CONTROL, 0, 2, 2); /* venc clock mode = normal */ |
| #endif |
| |
| r = request_irq(INT_24XX_DSS_IRQ, |
| cpu_is_omap24xx() |
| ? dss_irq_handler_omap2 |
| : dss_irq_handler_omap3, |
| 0, "OMAP DSS", NULL); |
| |
| if (r < 0) { |
| DSSERR("omap2 dss: request_irq failed\n"); |
| goto fail1; |
| } |
| |
| if (cpu_is_omap34xx()) { |
| dss.dpll4_m4_ck = clk_get(NULL, "dpll4_m4_ck"); |
| if (IS_ERR(dss.dpll4_m4_ck)) { |
| DSSERR("Failed to get dpll4_m4_ck\n"); |
| r = PTR_ERR(dss.dpll4_m4_ck); |
| goto fail2; |
| } |
| } |
| |
| dss.dsi_clk_source = DSS_SRC_DSS1_ALWON_FCLK; |
| dss.dispc_clk_source = DSS_SRC_DSS1_ALWON_FCLK; |
| |
| dss_save_context(); |
| |
| rev = dss_read_reg(DSS_REVISION); |
| printk(KERN_INFO "OMAP DSS rev %d.%d\n", |
| FLD_GET(rev, 7, 4), FLD_GET(rev, 3, 0)); |
| |
| return 0; |
| |
| fail2: |
| free_irq(INT_24XX_DSS_IRQ, NULL); |
| fail1: |
| iounmap(dss.base); |
| fail0: |
| return r; |
| } |
| |
| void dss_exit(void) |
| { |
| if (cpu_is_omap34xx()) |
| clk_put(dss.dpll4_m4_ck); |
| |
| free_irq(INT_24XX_DSS_IRQ, NULL); |
| |
| iounmap(dss.base); |
| } |
| |