blob: e569a0b32c70d20648aaf9023a5151fe621769e9 [file] [log] [blame]
/* Copyright (c) 2013-2014, 2016-2018, The Linux Foundation. All rights reserved.
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 and
* only version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
*/
#include <linux/bitops.h>
#include <linux/iopoll.h>
#include "mdp3.h"
#include "mdp3_dma.h"
#include "mdp3_hwio.h"
#include "mdss_debug.h"
#include "mdp3_ctrl.h"
#define DMA_STOP_POLL_SLEEP_US 1000
#define DMA_STOP_POLL_TIMEOUT_US 200000
#define DMA_HISTO_RESET_TIMEOUT_MS 40
#define DMA_LUT_CONFIG_MASK 0xfffffbe8
#define DMA_CCS_CONFIG_MASK 0xfffffc17
#define HIST_WAIT_TIMEOUT(frame) ((75 * HZ * (frame)) / 1000)
#define VSYNC_SELECT 0x024
#define VSYNC_TOTAL_LINES_SHIFT 21
#define VSYNC_COUNT_MASK 0x7ffff
#define VSYNC_THRESH_CONT_SHIFT 16
static void mdp3_vsync_intr_handler(int type, void *arg)
{
struct mdp3_dma *dma = (struct mdp3_dma *)arg;
struct mdp3_notification vsync_client;
struct mdp3_notification retire_client;
unsigned int wait_for_next_vs;
if (!dma) {
pr_err("dma is null\n");
return;
}
pr_debug("mdp3_vsync_intr_handler\n");
MDSS_XLOG(0x111, dma->vsync_period);
spin_lock(&dma->dma_lock);
vsync_client = dma->vsync_client;
retire_client = dma->retire_client;
wait_for_next_vs = !dma->vsync_status;
dma->vsync_status = 0;
if (wait_for_next_vs)
complete(&dma->vsync_comp);
spin_unlock(&dma->dma_lock);
if (vsync_client.handler) {
vsync_client.handler(vsync_client.arg);
} else {
if (wait_for_next_vs)
mdp3_irq_disable_nosync(type);
}
if (retire_client.handler)
retire_client.handler(retire_client.arg);
}
static void mdp3_dma_done_intr_handler(int type, void *arg)
{
struct mdp3_dma *dma = (struct mdp3_dma *)arg;
struct mdp3_notification dma_client;
if (!dma) {
pr_err("dma is null\n");
return;
}
pr_debug("mdp3_dma_done_intr_handler\n");
spin_lock(&dma->dma_lock);
dma_client = dma->dma_notifier_client;
complete(&dma->dma_comp);
spin_unlock(&dma->dma_lock);
mdp3_irq_disable_nosync(type);
if (dma_client.handler)
dma_client.handler(dma_client.arg);
}
static void mdp3_hist_done_intr_handler(int type, void *arg)
{
struct mdp3_dma *dma = (struct mdp3_dma *)arg;
u32 isr, mask;
if (!dma) {
pr_err("dma is null\n");
return;
}
isr = MDP3_REG_READ(MDP3_REG_DMA_P_HIST_INTR_STATUS);
mask = MDP3_REG_READ(MDP3_REG_DMA_P_HIST_INTR_ENABLE);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_INTR_CLEAR, isr);
isr &= mask;
if (isr == 0)
return;
if (isr & MDP3_DMA_P_HIST_INTR_HIST_DONE_BIT) {
spin_lock(&dma->histo_lock);
dma->histo_state = MDP3_DMA_HISTO_STATE_READY;
complete(&dma->histo_comp);
spin_unlock(&dma->histo_lock);
mdp3_hist_intr_notify(dma);
}
if (isr & MDP3_DMA_P_HIST_INTR_RESET_DONE_BIT) {
spin_lock(&dma->histo_lock);
dma->histo_state = MDP3_DMA_HISTO_STATE_IDLE;
complete(&dma->histo_comp);
spin_unlock(&dma->histo_lock);
}
}
void mdp3_dma_callback_enable(struct mdp3_dma *dma, int type)
{
int irq_bit;
pr_debug("mdp3_dma_callback_enable type=%d\n", type);
if (dma->dma_sel == MDP3_DMA_P) {
if (type & MDP3_DMA_CALLBACK_TYPE_HIST_RESET_DONE)
mdp3_irq_enable(MDP3_INTR_DMA_P_HISTO);
if (type & MDP3_DMA_CALLBACK_TYPE_HIST_DONE)
mdp3_irq_enable(MDP3_INTR_DMA_P_HISTO);
}
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO ||
dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_LCDC ||
dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_SPI_CMD) {
if (type & MDP3_DMA_CALLBACK_TYPE_VSYNC)
mdp3_irq_enable(MDP3_INTR_LCDC_START_OF_FRAME);
} else if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
if (type & MDP3_DMA_CALLBACK_TYPE_VSYNC) {
irq_bit = MDP3_INTR_SYNC_PRIMARY_LINE;
irq_bit += dma->dma_sel;
mdp3_irq_enable(irq_bit);
}
if (type & MDP3_DMA_CALLBACK_TYPE_DMA_DONE) {
irq_bit = MDP3_INTR_DMA_P_DONE;
if (dma->dma_sel == MDP3_DMA_S)
irq_bit = MDP3_INTR_DMA_S_DONE;
mdp3_irq_enable(irq_bit);
}
} else {
pr_err("mdp3_dma_callback_enable not supported interface\n");
}
}
void mdp3_dma_callback_disable(struct mdp3_dma *dma, int type)
{
int irq_bit;
pr_debug("mdp3_dma_callback_disable type=%d\n", type);
if (dma->dma_sel == MDP3_DMA_P) {
if (type & MDP3_DMA_CALLBACK_TYPE_HIST_RESET_DONE)
mdp3_irq_disable(MDP3_INTR_DMA_P_HISTO);
if (type & MDP3_DMA_CALLBACK_TYPE_HIST_DONE)
mdp3_irq_disable(MDP3_INTR_DMA_P_HISTO);
}
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO ||
dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_LCDC ||
dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_SPI_CMD) {
if (type & MDP3_DMA_CALLBACK_TYPE_VSYNC)
mdp3_irq_disable(MDP3_INTR_LCDC_START_OF_FRAME);
} else if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
if (type & MDP3_DMA_CALLBACK_TYPE_VSYNC) {
irq_bit = MDP3_INTR_SYNC_PRIMARY_LINE;
irq_bit += dma->dma_sel;
mdp3_irq_disable(irq_bit);
/*
* Clear read pointer interrupt before disabling clocks.
* Else pending ISR handling will result in NOC error
* since the clock will be disable after this point.
*/
mdp3_clear_irq(irq_bit);
}
if (type & MDP3_DMA_CALLBACK_TYPE_DMA_DONE) {
irq_bit = MDP3_INTR_DMA_P_DONE;
if (dma->dma_sel == MDP3_DMA_S)
irq_bit = MDP3_INTR_DMA_S_DONE;
mdp3_irq_disable(irq_bit);
}
}
}
static int mdp3_dma_callback_setup(struct mdp3_dma *dma)
{
int rc = 0;
struct mdp3_intr_cb vsync_cb = {
.cb = mdp3_vsync_intr_handler,
.data = dma,
};
struct mdp3_intr_cb dma_cb = {
.cb = mdp3_dma_done_intr_handler,
.data = dma,
};
struct mdp3_intr_cb hist_cb = {
.cb = mdp3_hist_done_intr_handler,
.data = dma,
};
if (dma->dma_sel == MDP3_DMA_P)
rc = mdp3_set_intr_callback(MDP3_INTR_DMA_P_HISTO, &hist_cb);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO ||
dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_LCDC)
rc |= mdp3_set_intr_callback(MDP3_INTR_LCDC_START_OF_FRAME,
&vsync_cb);
else if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
int irq_bit = MDP3_INTR_SYNC_PRIMARY_LINE;
irq_bit += dma->dma_sel;
rc |= mdp3_set_intr_callback(irq_bit, &vsync_cb);
irq_bit = MDP3_INTR_DMA_P_DONE;
if (dma->dma_sel == MDP3_DMA_S)
irq_bit = MDP3_INTR_DMA_S_DONE;
rc |= mdp3_set_intr_callback(irq_bit, &dma_cb);
} else {
pr_err("mdp3_dma_callback_setup not supported interface\n");
rc = -ENODEV;
}
return rc;
}
static void mdp3_dma_vsync_enable(struct mdp3_dma *dma,
struct mdp3_notification *vsync_client)
{
unsigned long flag;
int updated = 0;
int cb_type = MDP3_DMA_CALLBACK_TYPE_VSYNC;
pr_debug("mdp3_dma_vsync_enable\n");
spin_lock_irqsave(&dma->dma_lock, flag);
if (vsync_client) {
if (dma->vsync_client.handler != vsync_client->handler) {
dma->vsync_client = *vsync_client;
updated = 1;
}
} else {
if (dma->vsync_client.handler) {
dma->vsync_client.handler = NULL;
dma->vsync_client.arg = NULL;
updated = 1;
}
}
spin_unlock_irqrestore(&dma->dma_lock, flag);
if (updated) {
if (vsync_client && vsync_client->handler)
mdp3_dma_callback_enable(dma, cb_type);
else
mdp3_dma_callback_disable(dma, cb_type);
}
}
static void mdp3_dma_done_notifier(struct mdp3_dma *dma,
struct mdp3_notification *dma_client)
{
unsigned long flag;
spin_lock_irqsave(&dma->dma_lock, flag);
if (dma_client) {
dma->dma_notifier_client = *dma_client;
} else {
dma->dma_notifier_client.handler = NULL;
dma->dma_notifier_client.arg = NULL;
}
spin_unlock_irqrestore(&dma->dma_lock, flag);
}
int mdp3_dma_sync_config(struct mdp3_dma *dma,
struct mdp3_dma_source *source_config, struct mdp3_tear_check *te)
{
u32 vsync_clk_speed_hz, vclks_line, cfg;
int porch = source_config->vporch;
int height = source_config->height;
int total_lines = height + porch;
int dma_sel = dma->dma_sel;
vsync_clk_speed_hz = MDP_VSYNC_CLK_RATE;
cfg = te->sync_cfg_height << VSYNC_TOTAL_LINES_SHIFT;
total_lines *= te->frame_rate;
vclks_line = (total_lines) ? vsync_clk_speed_hz / total_lines : 0;
cfg |= BIT(19);
if (te->hw_vsync_mode)
cfg |= BIT(20);
if (te->refx100) {
vclks_line = vclks_line * te->frame_rate *
100 / te->refx100;
} else {
pr_warn("refx100 cannot be zero! Use 6000 as default\n");
vclks_line = vclks_line * te->frame_rate *
100 / 6000;
}
cfg |= (vclks_line & VSYNC_COUNT_MASK);
MDP3_REG_WRITE(MDP3_REG_SYNC_CONFIG_0 + dma_sel, cfg);
MDP3_REG_WRITE(MDP3_REG_VSYNC_SEL, VSYNC_SELECT);
MDP3_REG_WRITE(MDP3_REG_PRIMARY_VSYNC_INIT_VAL + dma_sel,
te->vsync_init_val);
MDP3_REG_WRITE(MDP3_REG_PRIMARY_RD_PTR_IRQ, te->rd_ptr_irq);
MDP3_REG_WRITE(MDP3_REG_SYNC_THRESH_0 + dma_sel,
((te->sync_threshold_continue << VSYNC_THRESH_CONT_SHIFT) |
te->sync_threshold_start));
MDP3_REG_WRITE(MDP3_REG_PRIMARY_START_P0S + dma_sel, te->start_pos);
MDP3_REG_WRITE(MDP3_REG_TEAR_CHECK_EN, te->tear_check_en);
return 0;
}
static int mdp3_dmap_config(struct mdp3_dma *dma,
struct mdp3_dma_source *source_config,
struct mdp3_dma_output_config *output_config,
bool splash_screen_active)
{
u32 dma_p_cfg_reg, dma_p_size, dma_p_out_xy;
dma_p_cfg_reg = source_config->format << 25;
if (output_config->dither_en)
dma_p_cfg_reg |= BIT(24);
dma_p_cfg_reg |= output_config->out_sel << 19;
dma_p_cfg_reg |= output_config->bit_mask_polarity << 18;
dma_p_cfg_reg |= output_config->color_components_flip << 14;
dma_p_cfg_reg |= output_config->pack_pattern << 8;
dma_p_cfg_reg |= output_config->pack_align << 7;
dma_p_cfg_reg |= output_config->color_comp_out_bits;
dma_p_size = source_config->width | (source_config->height << 16);
dma_p_out_xy = source_config->x | (source_config->y << 16);
if (!splash_screen_active) {
MDP3_REG_WRITE(MDP3_REG_DMA_P_CONFIG, dma_p_cfg_reg);
MDP3_REG_WRITE(MDP3_REG_DMA_P_SIZE, dma_p_size);
MDP3_REG_WRITE(MDP3_REG_DMA_P_IBUF_ADDR,
(u32)source_config->buf);
MDP3_REG_WRITE(MDP3_REG_DMA_P_IBUF_Y_STRIDE,
source_config->stride);
MDP3_REG_WRITE(MDP3_REG_DMA_P_OUT_XY, dma_p_out_xy);
MDP3_REG_WRITE(MDP3_REG_DMA_P_FETCH_CFG, 0x40);
}
dma->source_config = *source_config;
dma->output_config = *output_config;
if (dma->output_config.out_sel != MDP3_DMA_OUTPUT_SEL_DSI_CMD)
mdp3_irq_enable(MDP3_INTR_LCDC_UNDERFLOW);
mdp3_dma_callback_setup(dma);
return 0;
}
static void mdp3_dmap_config_source(struct mdp3_dma *dma)
{
struct mdp3_dma_source *source_config = &dma->source_config;
u32 dma_p_cfg_reg, dma_p_size;
dma_p_cfg_reg = MDP3_REG_READ(MDP3_REG_DMA_P_CONFIG);
dma_p_cfg_reg &= ~MDP3_DMA_IBUF_FORMAT_MASK;
dma_p_cfg_reg |= source_config->format << 25;
dma_p_cfg_reg &= ~MDP3_DMA_PACK_PATTERN_MASK;
dma_p_cfg_reg |= dma->output_config.pack_pattern << 8;
dma_p_size = dma->roi.w | (dma->roi.h << 16);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CONFIG, dma_p_cfg_reg);
MDP3_REG_WRITE(MDP3_REG_DMA_P_SIZE, dma_p_size);
MDP3_REG_WRITE(MDP3_REG_DMA_P_IBUF_Y_STRIDE, source_config->stride);
}
static int mdp3_dmas_config(struct mdp3_dma *dma,
struct mdp3_dma_source *source_config,
struct mdp3_dma_output_config *output_config,
bool splash_screen_active)
{
u32 dma_s_cfg_reg, dma_s_size, dma_s_out_xy;
dma_s_cfg_reg = source_config->format << 25;
if (output_config->dither_en)
dma_s_cfg_reg |= BIT(24);
dma_s_cfg_reg |= output_config->out_sel << 19;
dma_s_cfg_reg |= output_config->bit_mask_polarity << 18;
dma_s_cfg_reg |= output_config->color_components_flip << 14;
dma_s_cfg_reg |= output_config->pack_pattern << 8;
dma_s_cfg_reg |= output_config->pack_align << 7;
dma_s_cfg_reg |= output_config->color_comp_out_bits;
dma_s_size = source_config->width | (source_config->height << 16);
dma_s_out_xy = source_config->x | (source_config->y << 16);
if (!splash_screen_active) {
MDP3_REG_WRITE(MDP3_REG_DMA_S_CONFIG, dma_s_cfg_reg);
MDP3_REG_WRITE(MDP3_REG_DMA_S_SIZE, dma_s_size);
MDP3_REG_WRITE(MDP3_REG_DMA_S_IBUF_ADDR,
(u32)source_config->buf);
MDP3_REG_WRITE(MDP3_REG_DMA_S_IBUF_Y_STRIDE,
source_config->stride);
MDP3_REG_WRITE(MDP3_REG_DMA_S_OUT_XY, dma_s_out_xy);
MDP3_REG_WRITE(MDP3_REG_SECONDARY_RD_PTR_IRQ, 0x10);
}
dma->source_config = *source_config;
dma->output_config = *output_config;
mdp3_dma_callback_setup(dma);
return 0;
}
static void mdp3_dmas_config_source(struct mdp3_dma *dma)
{
struct mdp3_dma_source *source_config = &dma->source_config;
u32 dma_s_cfg_reg, dma_s_size;
dma_s_cfg_reg = MDP3_REG_READ(MDP3_REG_DMA_S_CONFIG);
dma_s_cfg_reg &= ~MDP3_DMA_IBUF_FORMAT_MASK;
dma_s_cfg_reg |= source_config->format << 25;
dma_s_size = source_config->width | (source_config->height << 16);
MDP3_REG_WRITE(MDP3_REG_DMA_S_CONFIG, dma_s_cfg_reg);
MDP3_REG_WRITE(MDP3_REG_DMA_S_SIZE, dma_s_size);
MDP3_REG_WRITE(MDP3_REG_DMA_S_IBUF_Y_STRIDE, source_config->stride);
}
static int mdp3_dmap_cursor_config(struct mdp3_dma *dma,
struct mdp3_dma_cursor *cursor)
{
u32 cursor_size, cursor_pos, blend_param, trans_mask;
cursor_size = cursor->width | (cursor->height << 16);
cursor_pos = cursor->x | (cursor->y << 16);
trans_mask = 0;
if (cursor->blend_config.mode == MDP3_DMA_CURSOR_BLEND_CONSTANT_ALPHA) {
blend_param = cursor->blend_config.constant_alpha << 24;
} else if (cursor->blend_config.mode ==
MDP3_DMA_CURSOR_BLEND_COLOR_KEYING) {
blend_param = cursor->blend_config.transparent_color;
trans_mask = cursor->blend_config.transparency_mask;
} else {
blend_param = 0;
}
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_FORMAT, cursor->format);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_SIZE, cursor_size);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_BUF_ADDR, (u32)cursor->buf);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_POS, cursor_pos);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_BLEND_CONFIG,
cursor->blend_config.mode);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_BLEND_PARAM, blend_param);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_BLEND_TRANS_MASK, trans_mask);
dma->cursor = *cursor;
return 0;
}
static int mdp3_dmap_ccs_config_internal(struct mdp3_dma *dma,
struct mdp3_dma_color_correct_config *config,
struct mdp3_dma_ccs *ccs)
{
int i;
u32 addr;
if (!ccs)
return -EINVAL;
if (config->ccs_enable) {
addr = MDP3_REG_DMA_P_CSC_MV1;
if (config->ccs_sel)
addr = MDP3_REG_DMA_P_CSC_MV2;
for (i = 0; i < 9; i++) {
MDP3_REG_WRITE(addr, ccs->mv[i]);
addr += 4;
}
addr = MDP3_REG_DMA_P_CSC_PRE_BV1;
if (config->pre_bias_sel)
addr = MDP3_REG_DMA_P_CSC_PRE_BV2;
for (i = 0; i < 3; i++) {
MDP3_REG_WRITE(addr, ccs->pre_bv[i]);
addr += 4;
}
addr = MDP3_REG_DMA_P_CSC_POST_BV1;
if (config->post_bias_sel)
addr = MDP3_REG_DMA_P_CSC_POST_BV2;
for (i = 0; i < 3; i++) {
MDP3_REG_WRITE(addr, ccs->post_bv[i]);
addr += 4;
}
addr = MDP3_REG_DMA_P_CSC_PRE_LV1;
if (config->pre_limit_sel)
addr = MDP3_REG_DMA_P_CSC_PRE_LV2;
for (i = 0; i < 6; i++) {
MDP3_REG_WRITE(addr, ccs->pre_lv[i]);
addr += 4;
}
addr = MDP3_REG_DMA_P_CSC_POST_LV1;
if (config->post_limit_sel)
addr = MDP3_REG_DMA_P_CSC_POST_LV2;
for (i = 0; i < 6; i++) {
MDP3_REG_WRITE(addr, ccs->post_lv[i]);
addr += 4;
}
}
return 0;
}
static void mdp3_ccs_update(struct mdp3_dma *dma, bool from_kickoff)
{
u32 cc_config;
bool ccs_updated = false, lut_updated = false;
struct mdp3_dma_ccs ccs;
cc_config = MDP3_REG_READ(MDP3_REG_DMA_P_COLOR_CORRECT_CONFIG);
if (dma->ccs_config.ccs_dirty) {
cc_config &= DMA_CCS_CONFIG_MASK;
if (dma->ccs_config.ccs_enable)
cc_config |= BIT(3);
else
cc_config &= ~BIT(3);
cc_config |= dma->ccs_config.ccs_sel << 5;
cc_config |= dma->ccs_config.pre_bias_sel << 6;
cc_config |= dma->ccs_config.post_bias_sel << 7;
cc_config |= dma->ccs_config.pre_limit_sel << 8;
cc_config |= dma->ccs_config.post_limit_sel << 9;
/*
* CCS dirty flag should be reset when call is made from frame
* kickoff, or else upon resume the flag would be dirty and LUT
* config could call this function thereby causing no register
* programming for CCS, which will cause screen to go dark
*/
if (from_kickoff)
dma->ccs_config.ccs_dirty = false;
ccs_updated = true;
}
if (dma->lut_config.lut_dirty) {
cc_config &= DMA_LUT_CONFIG_MASK;
cc_config |= dma->lut_config.lut_enable;
cc_config |= dma->lut_config.lut_position << 4;
cc_config |= dma->lut_config.lut_sel << 10;
dma->lut_config.lut_dirty = false;
lut_updated = true;
}
if (ccs_updated && from_kickoff) {
ccs.mv = dma->ccs_cache.csc_data.csc_mv;
ccs.pre_bv = dma->ccs_cache.csc_data.csc_pre_bv;
ccs.post_bv = dma->ccs_cache.csc_data.csc_post_bv;
ccs.pre_lv = dma->ccs_cache.csc_data.csc_pre_lv;
ccs.post_lv = dma->ccs_cache.csc_data.csc_post_lv;
mdp3_dmap_ccs_config_internal(dma, &dma->ccs_config, &ccs);
}
if (lut_updated || ccs_updated) {
MDP3_REG_WRITE(MDP3_REG_DMA_P_COLOR_CORRECT_CONFIG, cc_config);
/*
* Make sure ccs configuration update is done before continuing
* with the DMA transfer
*/
wmb(); /* ensure write is finished before progressing */
}
}
static int mdp3_dmap_ccs_config(struct mdp3_dma *dma,
struct mdp3_dma_color_correct_config *config,
struct mdp3_dma_ccs *ccs)
{
mdp3_dmap_ccs_config_internal(dma, config, ccs);
dma->ccs_config = *config;
if (dma->output_config.out_sel != MDP3_DMA_OUTPUT_SEL_DSI_CMD)
mdp3_ccs_update(dma, false);
return 0;
}
static int mdp3_dmap_lut_config(struct mdp3_dma *dma,
struct mdp3_dma_lut_config *config,
struct fb_cmap *cmap)
{
u32 addr, color;
int i;
if (config->lut_enable && cmap) {
addr = MDP3_REG_DMA_P_CSC_LUT1;
if (config->lut_sel)
addr = MDP3_REG_DMA_P_CSC_LUT2;
for (i = 0; i < MDP_LUT_SIZE; i++) {
color = cmap->green[i] & 0xff;
color |= (cmap->red[i] & 0xff) << 8;
color |= (cmap->blue[i] & 0xff) << 16;
MDP3_REG_WRITE(addr, color);
addr += 4;
}
}
dma->lut_config = *config;
if (dma->output_config.out_sel != MDP3_DMA_OUTPUT_SEL_DSI_CMD)
mdp3_ccs_update(dma, false);
return 0;
}
static int mdp3_dmap_histo_config(struct mdp3_dma *dma,
struct mdp3_dma_histogram_config *histo_config)
{
unsigned long flag;
u32 histo_bit_mask = 0, histo_control = 0;
u32 histo_isr_mask = MDP3_DMA_P_HIST_INTR_HIST_DONE_BIT |
MDP3_DMA_P_HIST_INTR_RESET_DONE_BIT;
spin_lock_irqsave(&dma->histo_lock, flag);
if (histo_config->bit_mask_polarity)
histo_bit_mask = BIT(31);
histo_bit_mask |= histo_config->bit_mask;
if (histo_config->auto_clear_en)
histo_control = BIT(0);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_FRAME_CNT,
histo_config->frame_count);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_BIT_MASK, histo_bit_mask);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_CONTROL, histo_control);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_INTR_ENABLE, histo_isr_mask);
spin_unlock_irqrestore(&dma->histo_lock, flag);
dma->histogram_config = *histo_config;
return 0;
}
int dma_bpp(int format)
{
int bpp;
switch (format) {
case MDP3_DMA_IBUF_FORMAT_RGB888:
bpp = 3;
break;
case MDP3_DMA_IBUF_FORMAT_RGB565:
bpp = 2;
break;
case MDP3_DMA_IBUF_FORMAT_XRGB8888:
bpp = 4;
break;
default:
bpp = 0;
}
return bpp;
}
static int mdp3_dmap_update(struct mdp3_dma *dma, void *buf,
struct mdp3_intf *intf, void *data)
{
unsigned long flag;
int cb_type = MDP3_DMA_CALLBACK_TYPE_VSYNC;
struct mdss_panel_data *panel;
int rc = 0;
int retry_count = 2;
ATRACE_BEGIN(__func__);
pr_debug("mdp3_dmap_update\n");
MDSS_XLOG(XLOG_FUNC_ENTRY, __LINE__);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
cb_type = MDP3_DMA_CALLBACK_TYPE_DMA_DONE;
if (intf->active) {
ATRACE_BEGIN("mdp3_wait_for_dma_comp");
retry_dma_done:
rc = wait_for_completion_timeout(&dma->dma_comp,
dma_timeout_value(dma));
if (rc <= 0 && --retry_count) {
int vsync_status;
vsync_status = (1 << MDP3_INTR_DMA_P_DONE) &
MDP3_REG_READ(MDP3_REG_INTR_STATUS);
if (!vsync_status) {
pr_err("%s: cmd timeout retry cnt %d\n",
__func__, retry_count);
goto retry_dma_done;
}
rc = -1;
}
ATRACE_END("mdp3_wait_for_dma_comp");
}
}
if (dma->update_src_cfg) {
if (dma->output_config.out_sel ==
MDP3_DMA_OUTPUT_SEL_DSI_VIDEO && intf->active)
pr_err("configuring dma source while it is active\n");
dma->dma_config_source(dma);
if (data) {
panel = (struct mdss_panel_data *)data;
if (panel->event_handler) {
panel->event_handler(panel,
MDSS_EVENT_ENABLE_PARTIAL_ROI, NULL);
panel->event_handler(panel,
MDSS_EVENT_DSI_STREAM_SIZE, NULL);
}
}
dma->update_src_cfg = false;
}
mutex_lock(&dma->pp_lock);
if (dma->ccs_config.ccs_dirty)
mdp3_ccs_update(dma, true);
mutex_unlock(&dma->pp_lock);
spin_lock_irqsave(&dma->dma_lock, flag);
MDP3_REG_WRITE(MDP3_REG_DMA_P_IBUF_ADDR, (u32)(buf +
dma->roi.y * dma->source_config.stride +
dma->roi.x * dma_bpp(dma->source_config.format)));
dma->source_config.buf = (int)buf;
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD)
MDP3_REG_WRITE(MDP3_REG_DMA_P_START, 1);
if (!intf->active) {
pr_debug("%s start interface\n", __func__);
intf->start(intf);
}
mb(); /* make sure everything is written before enable */
dma->vsync_status = MDP3_REG_READ(MDP3_REG_INTR_STATUS) &
(1 << MDP3_INTR_LCDC_START_OF_FRAME);
init_completion(&dma->vsync_comp);
spin_unlock_irqrestore(&dma->dma_lock, flag);
mdp3_dma_callback_enable(dma, cb_type);
pr_debug("%s wait for vsync_comp\n", __func__);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO) {
ATRACE_BEGIN("mdp3_wait_for_vsync_comp");
retry_vsync:
rc = wait_for_completion_timeout(&dma->vsync_comp,
dma_timeout_value(dma));
if (rc <= 0 && --retry_count) {
int vsync = MDP3_REG_READ(MDP3_REG_INTR_STATUS) &
(1 << MDP3_INTR_LCDC_START_OF_FRAME);
if (!vsync) {
pr_err("%s trying again count = %d\n",
__func__, retry_count);
goto retry_vsync;
}
rc = -1;
}
ATRACE_END("mdp3_wait_for_vsync_comp");
}
pr_debug("$%s wait for vsync_comp out\n", __func__);
ATRACE_END(__func__);
return rc;
}
static int mdp3_dmas_update(struct mdp3_dma *dma, void *buf,
struct mdp3_intf *intf, void *data)
{
unsigned long flag;
int cb_type = MDP3_DMA_CALLBACK_TYPE_VSYNC;
MDSS_XLOG(XLOG_FUNC_ENTRY, __LINE__);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
cb_type = MDP3_DMA_CALLBACK_TYPE_DMA_DONE;
if (intf->active)
wait_for_completion_killable(&dma->dma_comp);
}
spin_lock_irqsave(&dma->dma_lock, flag);
MDP3_REG_WRITE(MDP3_REG_DMA_S_IBUF_ADDR, (u32)buf);
dma->source_config.buf = (int)buf;
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD)
MDP3_REG_WRITE(MDP3_REG_DMA_S_START, 1);
if (!intf->active) {
pr_debug("mdp3_dmap_update start interface\n");
intf->start(intf);
}
wmb(); /* ensure write is finished before progressing */
init_completion(&dma->vsync_comp);
spin_unlock_irqrestore(&dma->dma_lock, flag);
mdp3_dma_callback_enable(dma, cb_type);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO)
wait_for_completion_killable(&dma->vsync_comp);
return 0;
}
static int mdp3_dmap_cursor_update(struct mdp3_dma *dma, int x, int y)
{
u32 cursor_pos;
cursor_pos = x | (y << 16);
MDP3_REG_WRITE(MDP3_REG_DMA_P_CURSOR_POS, cursor_pos);
dma->cursor.x = x;
dma->cursor.y = y;
return 0;
}
static int mdp3_dmap_histo_get(struct mdp3_dma *dma)
{
int i, state, timeout, ret;
u32 addr;
unsigned long flag;
spin_lock_irqsave(&dma->histo_lock, flag);
state = dma->histo_state;
spin_unlock_irqrestore(&dma->histo_lock, flag);
if (state != MDP3_DMA_HISTO_STATE_START &&
state != MDP3_DMA_HISTO_STATE_READY) {
pr_err("mdp3_dmap_histo_get invalid state %d\n", state);
return -EINVAL;
}
timeout = HIST_WAIT_TIMEOUT(dma->histogram_config.frame_count);
ret = wait_for_completion_killable_timeout(&dma->histo_comp, timeout);
if (ret == 0) {
pr_debug("mdp3_dmap_histo_get time out\n");
ret = -ETIMEDOUT;
} else if (ret < 0) {
pr_err("mdp3_dmap_histo_get interrupted\n");
}
if (ret < 0)
return ret;
if (dma->histo_state != MDP3_DMA_HISTO_STATE_READY) {
pr_debug("mdp3_dmap_histo_get after dma shut down\n");
return -EPERM;
}
addr = MDP3_REG_DMA_P_HIST_R_DATA;
for (i = 0; i < MDP_HISTOGRAM_BIN_NUM; i++) {
dma->histo_data.r_data[i] = MDP3_REG_READ(addr);
addr += 4;
}
addr = MDP3_REG_DMA_P_HIST_G_DATA;
for (i = 0; i < MDP_HISTOGRAM_BIN_NUM; i++) {
dma->histo_data.g_data[i] = MDP3_REG_READ(addr);
addr += 4;
}
addr = MDP3_REG_DMA_P_HIST_B_DATA;
for (i = 0; i < MDP_HISTOGRAM_BIN_NUM; i++) {
dma->histo_data.b_data[i] = MDP3_REG_READ(addr);
addr += 4;
}
dma->histo_data.extra[0] =
MDP3_REG_READ(MDP3_REG_DMA_P_HIST_EXTRA_INFO_0);
dma->histo_data.extra[1] =
MDP3_REG_READ(MDP3_REG_DMA_P_HIST_EXTRA_INFO_1);
spin_lock_irqsave(&dma->histo_lock, flag);
init_completion(&dma->histo_comp);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_START, 1);
wmb(); /* ensure write is finished before progressing */
dma->histo_state = MDP3_DMA_HISTO_STATE_START;
spin_unlock_irqrestore(&dma->histo_lock, flag);
return 0;
}
static int mdp3_dmap_histo_start(struct mdp3_dma *dma)
{
unsigned long flag;
if (dma->histo_state != MDP3_DMA_HISTO_STATE_IDLE)
return -EINVAL;
spin_lock_irqsave(&dma->histo_lock, flag);
init_completion(&dma->histo_comp);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_START, 1);
wmb(); /* ensure write is finished before progressing */
dma->histo_state = MDP3_DMA_HISTO_STATE_START;
spin_unlock_irqrestore(&dma->histo_lock, flag);
mdp3_dma_callback_enable(dma, MDP3_DMA_CALLBACK_TYPE_HIST_DONE);
return 0;
}
static int mdp3_dmap_histo_reset(struct mdp3_dma *dma)
{
unsigned long flag;
int ret;
spin_lock_irqsave(&dma->histo_lock, flag);
init_completion(&dma->histo_comp);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_INTR_ENABLE, BIT(0)|BIT(1));
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_RESET_SEQ_START, 1);
wmb(); /* ensure write is finished before progressing */
dma->histo_state = MDP3_DMA_HISTO_STATE_RESET;
spin_unlock_irqrestore(&dma->histo_lock, flag);
mdp3_dma_callback_enable(dma, MDP3_DMA_CALLBACK_TYPE_HIST_RESET_DONE);
ret = wait_for_completion_killable_timeout(&dma->histo_comp,
msecs_to_jiffies(DMA_HISTO_RESET_TIMEOUT_MS));
if (ret == 0) {
pr_err("mdp3_dmap_histo_reset time out\n");
ret = -ETIMEDOUT;
} else if (ret < 0) {
pr_err("mdp3_dmap_histo_reset interrupted\n");
} else {
ret = 0;
}
mdp3_dma_callback_disable(dma, MDP3_DMA_CALLBACK_TYPE_HIST_RESET_DONE);
return ret;
}
static int mdp3_dmap_histo_stop(struct mdp3_dma *dma)
{
unsigned long flag;
int cb_type = MDP3_DMA_CALLBACK_TYPE_HIST_RESET_DONE |
MDP3_DMA_CALLBACK_TYPE_HIST_DONE;
spin_lock_irqsave(&dma->histo_lock, flag);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_CANCEL_REQ, 1);
MDP3_REG_WRITE(MDP3_REG_DMA_P_HIST_INTR_ENABLE, 0);
wmb(); /* ensure write is finished before progressing */
dma->histo_state = MDP3_DMA_HISTO_STATE_IDLE;
complete(&dma->histo_comp);
spin_unlock_irqrestore(&dma->histo_lock, flag);
mdp3_dma_callback_disable(dma, cb_type);
return 0;
}
static int mdp3_dmap_histo_op(struct mdp3_dma *dma, u32 op)
{
int ret;
switch (op) {
case MDP3_DMA_HISTO_OP_START:
ret = mdp3_dmap_histo_start(dma);
break;
case MDP3_DMA_HISTO_OP_STOP:
case MDP3_DMA_HISTO_OP_CANCEL:
ret = mdp3_dmap_histo_stop(dma);
break;
case MDP3_DMA_HISTO_OP_RESET:
ret = mdp3_dmap_histo_reset(dma);
break;
default:
ret = -EINVAL;
}
return ret;
}
bool mdp3_dmap_busy(void)
{
u32 val;
val = MDP3_REG_READ(MDP3_REG_DISPLAY_STATUS);
pr_err("%s DMAP Status %s\n", __func__,
(val & MDP3_DMA_P_BUSY_BIT) ? "BUSY":"IDLE");
MDSS_XLOG(XLOG_FUNC_ENTRY, __LINE__,
(val & MDP3_DMA_P_BUSY_BIT) ? 1:0);
return val & MDP3_DMA_P_BUSY_BIT;
}
/*
* During underrun DMA_P registers are reset. Reprogramming CSC to prevent
* black screen
*/
static void mdp3_dmap_underrun_worker(struct work_struct *work)
{
struct mdp3_dma *dma;
dma = container_of(work, struct mdp3_dma, underrun_work);
mutex_lock(&dma->pp_lock);
if (dma->ccs_config.ccs_enable && dma->ccs_config.ccs_dirty) {
dma->cc_vect_sel = (dma->cc_vect_sel + 1) % 2;
dma->ccs_config.ccs_sel = dma->cc_vect_sel;
dma->ccs_config.pre_limit_sel = dma->cc_vect_sel;
dma->ccs_config.post_limit_sel = dma->cc_vect_sel;
dma->ccs_config.pre_bias_sel = dma->cc_vect_sel;
dma->ccs_config.post_bias_sel = dma->cc_vect_sel;
mdp3_ccs_update(dma, true);
}
mutex_unlock(&dma->pp_lock);
}
static int mdp3_dma_start(struct mdp3_dma *dma, struct mdp3_intf *intf)
{
unsigned long flag;
int cb_type = MDP3_DMA_CALLBACK_TYPE_VSYNC;
u32 dma_start_offset = MDP3_REG_DMA_P_START;
if (dma->dma_sel == MDP3_DMA_P)
dma_start_offset = MDP3_REG_DMA_P_START;
else if (dma->dma_sel == MDP3_DMA_S)
dma_start_offset = MDP3_REG_DMA_S_START;
else
return -EINVAL;
spin_lock_irqsave(&dma->dma_lock, flag);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD) {
cb_type |= MDP3_DMA_CALLBACK_TYPE_DMA_DONE;
MDP3_REG_WRITE(dma_start_offset, 1);
}
intf->start(intf);
wmb(); /* ensure write is finished before progressing */
init_completion(&dma->vsync_comp);
spin_unlock_irqrestore(&dma->dma_lock, flag);
if (dma->dma_sel == MDP3_DMA_P && dma->has_panic_ctrl)
MDP3_REG_WRITE(MDP3_PANIC_ROBUST_CTRL, BIT(0));
mdp3_dma_callback_enable(dma, cb_type);
pr_debug("mdp3_dma_start wait for vsync_comp in\n");
wait_for_completion_killable(&dma->vsync_comp);
pr_debug("mdp3_dma_start wait for vsync_comp out\n");
return 0;
}
static int mdp3_dma_stop(struct mdp3_dma *dma, struct mdp3_intf *intf)
{
int ret = 0;
u32 status, display_status_bit;
if (dma->dma_sel == MDP3_DMA_P)
display_status_bit = BIT(6);
else if (dma->dma_sel == MDP3_DMA_S)
display_status_bit = BIT(7);
else
return -EINVAL;
if (dma->dma_sel == MDP3_DMA_P && dma->has_panic_ctrl)
MDP3_REG_WRITE(MDP3_PANIC_ROBUST_CTRL, 0);
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_VIDEO)
display_status_bit |= BIT(11);
intf->stop(intf);
ret = readl_poll_timeout((mdp3_res->mdp_base + MDP3_REG_DISPLAY_STATUS),
status,
((status & display_status_bit) == 0),
DMA_STOP_POLL_SLEEP_US,
DMA_STOP_POLL_TIMEOUT_US);
mdp3_dma_callback_disable(dma, MDP3_DMA_CALLBACK_TYPE_VSYNC |
MDP3_DMA_CALLBACK_TYPE_DMA_DONE);
mdp3_irq_disable(MDP3_INTR_LCDC_UNDERFLOW);
MDP3_REG_WRITE(MDP3_REG_INTR_ENABLE, 0);
MDP3_REG_WRITE(MDP3_REG_INTR_CLEAR, 0xfffffff);
reinit_completion(&dma->dma_comp);
dma->vsync_client.handler = NULL;
/*
* Interrupts are disabled.
* Check for blocked dma done interrupt.
* Flush items waiting for dma done interrupt.
*/
if (dma->output_config.out_sel == MDP3_DMA_OUTPUT_SEL_DSI_CMD &&
atomic_read(&dma->session->dma_done_cnt))
mdp3_flush_dma_done(dma->session);
return ret;
}
int mdp3_dma_init(struct mdp3_dma *dma)
{
int ret = 0;
pr_debug("mdp3_dma_init\n");
switch (dma->dma_sel) {
case MDP3_DMA_P:
dma->dma_config = mdp3_dmap_config;
dma->dma_sync_config = mdp3_dma_sync_config;
dma->dma_config_source = mdp3_dmap_config_source;
dma->config_cursor = mdp3_dmap_cursor_config;
dma->config_ccs = mdp3_dmap_ccs_config;
dma->config_histo = mdp3_dmap_histo_config;
dma->config_lut = mdp3_dmap_lut_config;
dma->update = mdp3_dmap_update;
dma->update_cursor = mdp3_dmap_cursor_update;
dma->get_histo = mdp3_dmap_histo_get;
dma->histo_op = mdp3_dmap_histo_op;
dma->vsync_enable = mdp3_dma_vsync_enable;
dma->dma_done_notifier = mdp3_dma_done_notifier;
dma->start = mdp3_dma_start;
dma->stop = mdp3_dma_stop;
dma->busy = mdp3_dmap_busy;
INIT_WORK(&dma->underrun_work, mdp3_dmap_underrun_worker);
break;
case MDP3_DMA_S:
dma->dma_config = mdp3_dmas_config;
dma->dma_sync_config = mdp3_dma_sync_config;
dma->dma_config_source = mdp3_dmas_config_source;
dma->config_cursor = NULL;
dma->config_ccs = NULL;
dma->config_histo = NULL;
dma->config_lut = NULL;
dma->update = mdp3_dmas_update;
dma->update_cursor = NULL;
dma->get_histo = NULL;
dma->histo_op = NULL;
dma->vsync_enable = mdp3_dma_vsync_enable;
dma->start = mdp3_dma_start;
dma->stop = mdp3_dma_stop;
break;
case MDP3_DMA_E:
default:
ret = -ENODEV;
break;
}
spin_lock_init(&dma->dma_lock);
spin_lock_init(&dma->histo_lock);
init_completion(&dma->vsync_comp);
init_completion(&dma->dma_comp);
init_completion(&dma->histo_comp);
dma->vsync_client.handler = NULL;
dma->vsync_client.arg = NULL;
dma->histo_state = MDP3_DMA_HISTO_STATE_IDLE;
dma->update_src_cfg = false;
memset(&dma->cursor, 0, sizeof(dma->cursor));
memset(&dma->ccs_config, 0, sizeof(dma->ccs_config));
memset(&dma->histogram_config, 0, sizeof(dma->histogram_config));
return ret;
}
int lcdc_config(struct mdp3_intf *intf, struct mdp3_intf_cfg *cfg)
{
u32 temp;
struct mdp3_video_intf_cfg *v = &cfg->video;
temp = v->hsync_pulse_width | (v->hsync_period << 16);
MDP3_REG_WRITE(MDP3_REG_LCDC_HSYNC_CTL, temp);
MDP3_REG_WRITE(MDP3_REG_LCDC_VSYNC_PERIOD, v->vsync_period);
MDP3_REG_WRITE(MDP3_REG_LCDC_VSYNC_PULSE_WIDTH, v->vsync_pulse_width);
temp = v->display_start_x | (v->display_end_x << 16);
MDP3_REG_WRITE(MDP3_REG_LCDC_DISPLAY_HCTL, temp);
MDP3_REG_WRITE(MDP3_REG_LCDC_DISPLAY_V_START, v->display_start_y);
MDP3_REG_WRITE(MDP3_REG_LCDC_DISPLAY_V_END, v->display_end_y);
temp = v->active_start_x | (v->active_end_x);
if (v->active_h_enable)
temp |= BIT(31);
MDP3_REG_WRITE(MDP3_REG_LCDC_ACTIVE_HCTL, temp);
MDP3_REG_WRITE(MDP3_REG_LCDC_ACTIVE_V_START, v->active_start_y);
MDP3_REG_WRITE(MDP3_REG_LCDC_ACTIVE_V_END, v->active_end_y);
MDP3_REG_WRITE(MDP3_REG_LCDC_HSYNC_SKEW, v->hsync_skew);
temp = 0;
if (!v->hsync_polarity)
temp = BIT(0);
if (!v->vsync_polarity)
temp = BIT(1);
if (!v->de_polarity)
temp = BIT(2);
MDP3_REG_WRITE(MDP3_REG_LCDC_CTL_POLARITY, temp);
return 0;
}
int lcdc_start(struct mdp3_intf *intf)
{
MDP3_REG_WRITE(MDP3_REG_LCDC_EN, BIT(0));
wmb(); /* ensure write is finished before progressing */
intf->active = true;
return 0;
}
int lcdc_stop(struct mdp3_intf *intf)
{
MDP3_REG_WRITE(MDP3_REG_LCDC_EN, 0);
wmb(); /* ensure write is finished before progressing */
intf->active = false;
return 0;
}
int dsi_video_config(struct mdp3_intf *intf, struct mdp3_intf_cfg *cfg)
{
u32 temp;
struct mdp3_video_intf_cfg *v = &cfg->video;
pr_debug("dsi_video_config\n");
temp = v->hsync_pulse_width | (v->hsync_period << 16);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_HSYNC_CTL, temp);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_VSYNC_PERIOD, v->vsync_period);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_VSYNC_PULSE_WIDTH,
v->vsync_pulse_width);
temp = v->display_start_x | (v->display_end_x << 16);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_DISPLAY_HCTL, temp);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_DISPLAY_V_START, v->display_start_y);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_DISPLAY_V_END, v->display_end_y);
temp = v->active_start_x | (v->active_end_x << 16);
if (v->active_h_enable)
temp |= BIT(31);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_ACTIVE_HCTL, temp);
temp = v->active_start_y;
if (v->active_v_enable)
temp |= BIT(31);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_ACTIVE_V_START, temp);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_ACTIVE_V_END, v->active_end_y);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_HSYNC_SKEW, v->hsync_skew);
temp = 0;
if (!v->hsync_polarity)
temp |= BIT(0);
if (!v->vsync_polarity)
temp |= BIT(1);
if (!v->de_polarity)
temp |= BIT(2);
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_CTL_POLARITY, temp);
v->underflow_color |= 0x80000000;
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_UNDERFLOW_CTL, v->underflow_color);
return 0;
}
int dsi_video_start(struct mdp3_intf *intf)
{
pr_debug("dsi_video_start\n");
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_EN, BIT(0));
wmb(); /* ensure write is finished before progressing */
intf->active = true;
return 0;
}
int dsi_video_stop(struct mdp3_intf *intf)
{
pr_debug("dsi_video_stop\n");
MDP3_REG_WRITE(MDP3_REG_DSI_VIDEO_EN, 0);
wmb(); /* ensure write is finished before progressing */
intf->active = false;
return 0;
}
int dsi_cmd_config(struct mdp3_intf *intf, struct mdp3_intf_cfg *cfg)
{
u32 id_map = 0;
u32 trigger_en = 0;
if (cfg->dsi_cmd.primary_dsi_cmd_id)
id_map = BIT(0);
if (cfg->dsi_cmd.secondary_dsi_cmd_id)
id_map = BIT(4);
if (cfg->dsi_cmd.dsi_cmd_tg_intf_sel)
trigger_en = BIT(4);
MDP3_REG_WRITE(MDP3_REG_DSI_CMD_MODE_ID_MAP, id_map);
MDP3_REG_WRITE(MDP3_REG_DSI_CMD_MODE_TRIGGER_EN, trigger_en);
return 0;
}
int dsi_cmd_start(struct mdp3_intf *intf)
{
intf->active = true;
return 0;
}
int dsi_cmd_stop(struct mdp3_intf *intf)
{
intf->active = false;
return 0;
}
static int spi_cmd_config(struct mdp3_intf *intf, struct mdp3_intf_cfg *cfg)
{
return 0;
}
static int spi_cmd_start(struct mdp3_intf *intf)
{
intf->active = true;
return 0;
}
static int spi_cmd_stop(struct mdp3_intf *intf)
{
intf->active = false;
return 0;
}
int mdp3_intf_init(struct mdp3_intf *intf)
{
switch (intf->cfg.type) {
case MDP3_DMA_OUTPUT_SEL_LCDC:
intf->config = lcdc_config;
intf->start = lcdc_start;
intf->stop = lcdc_stop;
break;
case MDP3_DMA_OUTPUT_SEL_DSI_VIDEO:
intf->config = dsi_video_config;
intf->start = dsi_video_start;
intf->stop = dsi_video_stop;
break;
case MDP3_DMA_OUTPUT_SEL_DSI_CMD:
intf->config = dsi_cmd_config;
intf->start = dsi_cmd_start;
intf->stop = dsi_cmd_stop;
break;
case MDP3_DMA_OUTPUT_SEL_SPI_CMD:
intf->config = spi_cmd_config;
intf->start = spi_cmd_start;
intf->stop = spi_cmd_stop;
break;
default:
return -EINVAL;
}
return 0;
}