| /* |
| * Copyright (c) 2017, The Linux Foundation. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 and |
| * only version 2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| */ |
| #include <linux/clk.h> |
| #include <linux/dmaengine.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/interrupt.h> |
| #include <linux/ipc_logging.h> |
| #include <linux/io.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/of_platform.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/qcom-geni-se.h> |
| #include <linux/msm_gpi.h> |
| #include <linux/spi/spi.h> |
| #include <linux/spi/spi-geni-qcom.h> |
| |
| #define SPI_NUM_CHIPSELECT (4) |
| #define SPI_XFER_TIMEOUT_MS (250) |
| /* SPI SE specific registers */ |
| #define SE_SPI_CPHA (0x224) |
| #define SE_SPI_LOOPBACK (0x22C) |
| #define SE_SPI_CPOL (0x230) |
| #define SE_SPI_DEMUX_OUTPUT_INV (0x24C) |
| #define SE_SPI_DEMUX_SEL (0x250) |
| #define SE_SPI_TRANS_CFG (0x25C) |
| #define SE_SPI_WORD_LEN (0x268) |
| #define SE_SPI_TX_TRANS_LEN (0x26C) |
| #define SE_SPI_RX_TRANS_LEN (0x270) |
| #define SE_SPI_PRE_POST_CMD_DLY (0x274) |
| #define SE_SPI_DELAY_COUNTERS (0x278) |
| |
| /* SE_SPI_CPHA register fields */ |
| #define CPHA (BIT(0)) |
| |
| /* SE_SPI_LOOPBACK register fields */ |
| #define LOOPBACK_ENABLE (0x1) |
| #define NORMAL_MODE (0x0) |
| #define LOOPBACK_MSK (GENMASK(1, 0)) |
| |
| /* SE_SPI_CPOL register fields */ |
| #define CPOL (BIT(2)) |
| |
| /* SE_SPI_DEMUX_OUTPUT_INV register fields */ |
| #define CS_DEMUX_OUTPUT_INV_MSK (GENMASK(3, 0)) |
| |
| /* SE_SPI_DEMUX_SEL register fields */ |
| #define CS_DEMUX_OUTPUT_SEL (GENMASK(3, 0)) |
| |
| /* SE_SPI_TX_TRANS_CFG register fields */ |
| #define CS_TOGGLE (BIT(0)) |
| |
| /* SE_SPI_WORD_LEN register fields */ |
| #define WORD_LEN_MSK (GENMASK(9, 0)) |
| #define MIN_WORD_LEN (4) |
| |
| /* SPI_TX/SPI_RX_TRANS_LEN fields */ |
| #define TRANS_LEN_MSK (GENMASK(23, 0)) |
| |
| /* SE_SPI_DELAY_COUNTERS */ |
| #define SPI_INTER_WORDS_DELAY_MSK (GENMASK(9, 0)) |
| #define SPI_CS_CLK_DELAY_MSK (GENMASK(19, 10)) |
| #define SPI_CS_CLK_DELAY_SHFT (10) |
| |
| /* M_CMD OP codes for SPI */ |
| #define SPI_TX_ONLY (1) |
| #define SPI_RX_ONLY (2) |
| #define SPI_FULL_DUPLEX (3) |
| #define SPI_TX_RX (7) |
| #define SPI_CS_ASSERT (8) |
| #define SPI_CS_DEASSERT (9) |
| #define SPI_SCK_ONLY (10) |
| /* M_CMD params for SPI */ |
| #define SPI_PRE_CMD_DELAY BIT(0) |
| #define TIMESTAMP_BEFORE BIT(1) |
| #define FRAGMENTATION BIT(2) |
| #define TIMESTAMP_AFTER BIT(3) |
| #define POST_CMD_DELAY BIT(4) |
| |
| #define SPI_CORE2X_VOTE (10000) |
| /* GSI CONFIG0 TRE Params */ |
| /* Flags bit fields */ |
| #define GSI_LOOPBACK_EN (BIT(0)) |
| #define GSI_CS_TOGGLE (BIT(3)) |
| #define GSI_CPHA (BIT(4)) |
| #define GSI_CPOL (BIT(5)) |
| |
| #define MAX_TX_SG (3) |
| #define NUM_SPI_XFER (8) |
| |
| struct gsi_desc_cb { |
| struct spi_master *spi; |
| struct spi_transfer *xfer; |
| }; |
| |
| struct spi_geni_gsi { |
| struct msm_gpi_tre config0_tre; |
| struct msm_gpi_tre go_tre; |
| struct msm_gpi_tre tx_dma_tre; |
| struct msm_gpi_tre rx_dma_tre; |
| struct scatterlist tx_sg[MAX_TX_SG]; |
| struct scatterlist rx_sg; |
| dma_cookie_t tx_cookie; |
| dma_cookie_t rx_cookie; |
| struct msm_gpi_dma_async_tx_cb_param tx_cb_param; |
| struct msm_gpi_dma_async_tx_cb_param rx_cb_param; |
| struct dma_async_tx_descriptor *tx_desc; |
| struct dma_async_tx_descriptor *rx_desc; |
| struct gsi_desc_cb desc_cb; |
| }; |
| |
| struct spi_geni_master { |
| struct se_geni_rsc spi_rsc; |
| resource_size_t phys_addr; |
| resource_size_t size; |
| void __iomem *base; |
| int irq; |
| struct device *dev; |
| int rx_fifo_depth; |
| int tx_fifo_depth; |
| int tx_fifo_width; |
| int tx_wm; |
| bool setup; |
| u32 cur_speed_hz; |
| int cur_word_len; |
| unsigned int tx_rem_bytes; |
| unsigned int rx_rem_bytes; |
| struct spi_transfer *cur_xfer; |
| struct completion xfer_done; |
| struct device *wrapper_dev; |
| int oversampling; |
| struct spi_geni_gsi *gsi; |
| struct dma_chan *tx; |
| struct dma_chan *rx; |
| struct msm_gpi_ctrl tx_event; |
| struct msm_gpi_ctrl rx_event; |
| struct completion tx_cb; |
| struct completion rx_cb; |
| bool qn_err; |
| int cur_xfer_mode; |
| int num_tx_eot; |
| int num_rx_eot; |
| int num_xfers; |
| void *ipc; |
| }; |
| |
| static struct spi_master *get_spi_master(struct device *dev) |
| { |
| struct platform_device *pdev = to_platform_device(dev); |
| struct spi_master *spi = platform_get_drvdata(pdev); |
| |
| return spi; |
| } |
| |
| static int get_spi_clk_cfg(u32 speed_hz, struct spi_geni_master *mas, |
| int *clk_idx, int *clk_div) |
| { |
| unsigned long sclk_freq; |
| struct se_geni_rsc *rsc = &mas->spi_rsc; |
| int ret = 0; |
| |
| ret = geni_se_clk_freq_match(&mas->spi_rsc, |
| (speed_hz * mas->oversampling), clk_idx, |
| &sclk_freq, true); |
| if (ret) { |
| dev_err(mas->dev, "%s: Failed(%d) to find src clk for 0x%x\n", |
| __func__, ret, speed_hz); |
| return ret; |
| } |
| |
| *clk_div = ((sclk_freq / mas->oversampling) / speed_hz); |
| if (!(*clk_div)) { |
| dev_err(mas->dev, "%s:Err:sclk:%lu oversampling:%d speed:%u\n", |
| __func__, sclk_freq, mas->oversampling, speed_hz); |
| return -EINVAL; |
| } |
| |
| dev_dbg(mas->dev, "%s: req %u sclk %lu, idx %d, div %d\n", __func__, |
| speed_hz, sclk_freq, *clk_idx, *clk_div); |
| ret = clk_set_rate(rsc->se_clk, sclk_freq); |
| if (ret) |
| dev_err(mas->dev, "%s: clk_set_rate failed %d\n", |
| __func__, ret); |
| return ret; |
| } |
| |
| static void spi_setup_word_len(struct spi_geni_master *mas, u32 mode, |
| int bits_per_word) |
| { |
| int pack_words = 1; |
| bool msb_first = (mode & SPI_LSB_FIRST) ? false : true; |
| u32 word_len = geni_read_reg(mas->base, SE_SPI_WORD_LEN); |
| unsigned long cfg0, cfg1; |
| |
| /* |
| * If bits_per_word isn't a byte aligned value, set the packing to be |
| * 1 SPI word per FIFO word. |
| */ |
| if (!(mas->tx_fifo_width % bits_per_word)) |
| pack_words = mas->tx_fifo_width / bits_per_word; |
| word_len &= ~WORD_LEN_MSK; |
| word_len |= ((bits_per_word - MIN_WORD_LEN) & WORD_LEN_MSK); |
| se_config_packing(mas->base, bits_per_word, pack_words, msb_first); |
| geni_write_reg(word_len, mas->base, SE_SPI_WORD_LEN); |
| se_get_packing_config(bits_per_word, pack_words, msb_first, |
| &cfg0, &cfg1); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: cfg0 %lu cfg1 %lu bpw %d pack_words %d\n", __func__, |
| cfg0, cfg1, bits_per_word, pack_words); |
| } |
| |
| static int setup_fifo_params(struct spi_device *spi_slv, |
| struct spi_master *spi) |
| { |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| u16 mode = spi_slv->mode; |
| u32 loopback_cfg = geni_read_reg(mas->base, SE_SPI_LOOPBACK); |
| u32 cpol = geni_read_reg(mas->base, SE_SPI_CPOL); |
| u32 cpha = geni_read_reg(mas->base, SE_SPI_CPHA); |
| u32 demux_sel = 0; |
| u32 demux_output_inv = 0; |
| u32 clk_sel = geni_read_reg(mas->base, SE_GENI_CLK_SEL); |
| u32 m_clk_cfg = geni_read_reg(mas->base, GENI_SER_M_CLK_CFG); |
| int ret = 0; |
| int idx; |
| int div; |
| struct spi_geni_qcom_ctrl_data *delay_params = NULL; |
| u32 spi_delay_params = 0; |
| |
| loopback_cfg &= ~LOOPBACK_MSK; |
| cpol &= ~CPOL; |
| cpha &= ~CPHA; |
| |
| if (mode & SPI_LOOP) |
| loopback_cfg |= LOOPBACK_ENABLE; |
| |
| if (mode & SPI_CPOL) |
| cpol |= CPOL; |
| |
| if (mode & SPI_CPHA) |
| cpha |= CPHA; |
| |
| if (spi_slv->mode & SPI_CS_HIGH) |
| demux_output_inv |= BIT(spi_slv->chip_select); |
| |
| if (spi_slv->controller_data) { |
| u32 cs_clk_delay = 0; |
| u32 inter_words_delay = 0; |
| |
| delay_params = |
| (struct spi_geni_qcom_ctrl_data *) spi_slv->controller_data; |
| cs_clk_delay = |
| (delay_params->spi_cs_clk_delay << SPI_CS_CLK_DELAY_SHFT) |
| & SPI_CS_CLK_DELAY_MSK; |
| inter_words_delay = |
| delay_params->spi_inter_words_delay & |
| SPI_INTER_WORDS_DELAY_MSK; |
| spi_delay_params = |
| (inter_words_delay | cs_clk_delay); |
| } |
| |
| demux_sel = spi_slv->chip_select; |
| mas->cur_speed_hz = spi_slv->max_speed_hz; |
| mas->cur_word_len = spi_slv->bits_per_word; |
| |
| ret = get_spi_clk_cfg(mas->cur_speed_hz, mas, &idx, &div); |
| if (ret) { |
| dev_err(mas->dev, "Err setting clks ret(%d) for %d\n", |
| ret, mas->cur_speed_hz); |
| goto setup_fifo_params_exit; |
| } |
| |
| clk_sel |= (idx & CLK_SEL_MSK); |
| m_clk_cfg |= ((div << CLK_DIV_SHFT) | SER_CLK_EN); |
| spi_setup_word_len(mas, spi_slv->mode, spi_slv->bits_per_word); |
| geni_write_reg(loopback_cfg, mas->base, SE_SPI_LOOPBACK); |
| geni_write_reg(demux_sel, mas->base, SE_SPI_DEMUX_SEL); |
| geni_write_reg(cpha, mas->base, SE_SPI_CPHA); |
| geni_write_reg(cpol, mas->base, SE_SPI_CPOL); |
| geni_write_reg(demux_output_inv, mas->base, SE_SPI_DEMUX_OUTPUT_INV); |
| geni_write_reg(clk_sel, mas->base, SE_GENI_CLK_SEL); |
| geni_write_reg(m_clk_cfg, mas->base, GENI_SER_M_CLK_CFG); |
| geni_write_reg(spi_delay_params, mas->base, SE_SPI_DELAY_COUNTERS); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:Loopback%d demux_sel0x%x demux_op_inv 0x%x clk_cfg 0x%x\n", |
| __func__, loopback_cfg, demux_sel, demux_output_inv, m_clk_cfg); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:clk_sel 0x%x cpol %d cpha %d delay 0x%x\n", __func__, |
| clk_sel, cpol, cpha, spi_delay_params); |
| /* Ensure message level attributes are written before returning */ |
| mb(); |
| setup_fifo_params_exit: |
| return ret; |
| } |
| |
| |
| static int select_xfer_mode(struct spi_master *spi, |
| struct spi_message *spi_msg) |
| { |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| int mode = FIFO_MODE; |
| int fifo_disable = (geni_read_reg(mas->base, GENI_IF_FIFO_DISABLE_RO) & |
| FIFO_IF_DISABLE); |
| bool dma_chan_valid = |
| !(IS_ERR_OR_NULL(mas->tx) || IS_ERR_OR_NULL(mas->rx)); |
| |
| /* |
| * If FIFO Interface is disabled and there are no DMA channels then we |
| * can't do this transfer. |
| * If FIFO interface is disabled, we can do GSI only, |
| * else pick FIFO mode. |
| */ |
| if (fifo_disable && !dma_chan_valid) |
| mode = -EINVAL; |
| else if (dma_chan_valid) |
| mode = GSI_DMA; |
| else |
| mode = FIFO_MODE; |
| return mode; |
| } |
| |
| static struct msm_gpi_tre *setup_config0_tre(struct spi_transfer *xfer, |
| struct spi_geni_master *mas, u16 mode, |
| u32 cs_clk_delay, u32 inter_words_delay) |
| { |
| struct msm_gpi_tre *c0_tre = &mas->gsi[mas->num_xfers].config0_tre; |
| u8 flags = 0; |
| u8 word_len = 0; |
| u8 pack = 0; |
| int div = 0; |
| int idx = 0; |
| int ret = 0; |
| |
| if (IS_ERR_OR_NULL(c0_tre)) |
| return c0_tre; |
| |
| if (mode & SPI_LOOP) |
| flags |= GSI_LOOPBACK_EN; |
| |
| if (mode & SPI_CPOL) |
| flags |= GSI_CPOL; |
| |
| if (mode & SPI_CPHA) |
| flags |= GSI_CPHA; |
| |
| if (xfer->cs_change) |
| flags |= GSI_CS_TOGGLE; |
| |
| word_len = xfer->bits_per_word - MIN_WORD_LEN; |
| pack |= (GSI_TX_PACK_EN | GSI_RX_PACK_EN); |
| ret = get_spi_clk_cfg(mas->cur_speed_hz, mas, &idx, &div); |
| if (ret) { |
| dev_err(mas->dev, "%s:Err setting clks:%d\n", __func__, ret); |
| return ERR_PTR(ret); |
| } |
| c0_tre->dword[0] = MSM_GPI_SPI_CONFIG0_TRE_DWORD0(pack, flags, |
| word_len); |
| c0_tre->dword[1] = MSM_GPI_SPI_CONFIG0_TRE_DWORD1(0, cs_clk_delay, |
| inter_words_delay); |
| c0_tre->dword[2] = MSM_GPI_SPI_CONFIG0_TRE_DWORD2(idx, div); |
| c0_tre->dword[3] = MSM_GPI_SPI_CONFIG0_TRE_DWORD3(0, 0, 0, 1); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: flags 0x%x word %d pack %d idx %d div %d\n", |
| __func__, flags, word_len, pack, idx, div); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: cs_clk_delay %d inter_words_delay %d\n", __func__, |
| cs_clk_delay, inter_words_delay); |
| return c0_tre; |
| } |
| |
| static struct msm_gpi_tre *setup_go_tre(int cmd, int cs, int rx_len, int flags, |
| struct spi_geni_master *mas) |
| { |
| struct msm_gpi_tre *go_tre = &mas->gsi[mas->num_xfers].go_tre; |
| int chain; |
| int eot; |
| int eob; |
| |
| if (IS_ERR_OR_NULL(go_tre)) |
| return go_tre; |
| |
| go_tre->dword[0] = MSM_GPI_SPI_GO_TRE_DWORD0(flags, cs, cmd); |
| go_tre->dword[1] = MSM_GPI_SPI_GO_TRE_DWORD1; |
| go_tre->dword[2] = MSM_GPI_SPI_GO_TRE_DWORD2(rx_len); |
| if (cmd == SPI_RX_ONLY) { |
| eot = 0; |
| chain = 0; |
| eob = 1; |
| } else { |
| eot = 0; |
| chain = 1; |
| eob = 0; |
| } |
| go_tre->dword[3] = MSM_GPI_SPI_GO_TRE_DWORD3(0, eot, eob, chain); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: rx len %d flags 0x%x cs %d cmd %d eot %d eob %d chain %d\n", |
| __func__, rx_len, flags, cs, cmd, eot, eob, chain); |
| return go_tre; |
| } |
| |
| static struct msm_gpi_tre *setup_dma_tre(struct msm_gpi_tre *tre, |
| dma_addr_t buf, u32 len, |
| struct spi_geni_master *mas, |
| bool is_tx) |
| { |
| if (IS_ERR_OR_NULL(tre)) |
| return tre; |
| |
| tre->dword[0] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD0(buf); |
| tre->dword[1] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD1(buf); |
| tre->dword[2] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD2(len); |
| tre->dword[3] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD3(0, is_tx, 0, 0); |
| return tre; |
| } |
| |
| static void spi_gsi_ch_cb(struct dma_chan *ch, struct msm_gpi_cb const *cb, |
| void *ptr) |
| { |
| struct spi_master *spi = ptr; |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| |
| switch (cb->cb_event) { |
| case MSM_GPI_QUP_NOTIFY: |
| case MSM_GPI_QUP_MAX_EVENT: |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:cb_ev%d status%llu ts%llu count%llu\n", |
| __func__, cb->cb_event, cb->status, |
| cb->timestamp, cb->count); |
| break; |
| case MSM_GPI_QUP_ERROR: |
| case MSM_GPI_QUP_CH_ERROR: |
| case MSM_GPI_QUP_FW_ERROR: |
| case MSM_GPI_QUP_PENDING_EVENT: |
| case MSM_GPI_QUP_EOT_DESC_MISMATCH: |
| case MSM_GPI_QUP_SW_ERROR: |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: cb_ev %d status %llu ts %llu count %llu\n", |
| __func__, cb->cb_event, cb->status, |
| cb->timestamp, cb->count); |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "err.routine %u, err.type %u, err.code %u\n", |
| cb->error_log.routine, |
| cb->error_log.type, |
| cb->error_log.error_code); |
| mas->qn_err = true; |
| complete_all(&mas->tx_cb); |
| complete_all(&mas->rx_cb); |
| |
| break; |
| }; |
| } |
| |
| static void spi_gsi_rx_callback(void *cb) |
| { |
| struct msm_gpi_dma_async_tx_cb_param *cb_param = |
| (struct msm_gpi_dma_async_tx_cb_param *)cb; |
| struct gsi_desc_cb *desc_cb = (struct gsi_desc_cb *)cb_param->userdata; |
| struct spi_master *spi = desc_cb->spi; |
| struct spi_transfer *xfer = desc_cb->xfer; |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| |
| if (xfer->rx_buf) { |
| if (cb_param->status == MSM_GPI_TCE_UNEXP_ERR) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Unexpected GSI CB error\n", __func__); |
| return; |
| } |
| if (cb_param->length == xfer->len) { |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s\n", __func__); |
| complete(&mas->rx_cb); |
| } else { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Length mismatch. Expected %d Callback %d\n", |
| __func__, xfer->len, cb_param->length); |
| } |
| } |
| } |
| |
| static void spi_gsi_tx_callback(void *cb) |
| { |
| struct msm_gpi_dma_async_tx_cb_param *cb_param = cb; |
| struct gsi_desc_cb *desc_cb = (struct gsi_desc_cb *)cb_param->userdata; |
| struct spi_master *spi = desc_cb->spi; |
| struct spi_transfer *xfer = desc_cb->xfer; |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| |
| if (xfer->tx_buf) { |
| if (cb_param->status == MSM_GPI_TCE_UNEXP_ERR) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Unexpected GSI CB error\n", __func__); |
| return; |
| } |
| if (cb_param->length == xfer->len) { |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s\n", __func__); |
| complete(&mas->tx_cb); |
| } else { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Length mismatch. Expected %d Callback %d\n", |
| __func__, xfer->len, cb_param->length); |
| } |
| } |
| } |
| |
| static int setup_gsi_xfer(struct spi_transfer *xfer, |
| struct spi_geni_master *mas, |
| struct spi_device *spi_slv, |
| struct spi_master *spi) |
| { |
| int ret = 0; |
| struct msm_gpi_tre *c0_tre = NULL; |
| struct msm_gpi_tre *go_tre = NULL; |
| struct msm_gpi_tre *tx_tre = NULL; |
| struct msm_gpi_tre *rx_tre = NULL; |
| struct scatterlist *xfer_tx_sg = mas->gsi[mas->num_xfers].tx_sg; |
| struct scatterlist *xfer_rx_sg = &mas->gsi[mas->num_xfers].rx_sg; |
| int rx_nent = 0; |
| int tx_nent = 0; |
| u8 cmd = 0; |
| u8 cs = 0; |
| u32 rx_len = 0; |
| int go_flags = 0; |
| unsigned long flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK; |
| struct spi_geni_qcom_ctrl_data *delay_params = NULL; |
| u32 cs_clk_delay = 0; |
| u32 inter_words_delay = 0; |
| |
| if (spi_slv->controller_data) { |
| delay_params = |
| (struct spi_geni_qcom_ctrl_data *) spi_slv->controller_data; |
| |
| cs_clk_delay = |
| delay_params->spi_cs_clk_delay; |
| inter_words_delay = |
| delay_params->spi_inter_words_delay; |
| } |
| |
| if ((xfer->bits_per_word != mas->cur_word_len) || |
| (xfer->speed_hz != mas->cur_speed_hz)) { |
| mas->cur_word_len = xfer->bits_per_word; |
| mas->cur_speed_hz = xfer->speed_hz; |
| tx_nent++; |
| c0_tre = setup_config0_tre(xfer, mas, spi_slv->mode, |
| cs_clk_delay, inter_words_delay); |
| if (IS_ERR_OR_NULL(c0_tre)) { |
| dev_err(mas->dev, "%s:Err setting c0tre:%d\n", |
| __func__, ret); |
| return PTR_ERR(c0_tre); |
| } |
| } |
| |
| if (!(mas->cur_word_len % MIN_WORD_LEN)) { |
| rx_len = ((xfer->len << 3) / mas->cur_word_len); |
| } else { |
| int bytes_per_word = (mas->cur_word_len / BITS_PER_BYTE) + 1; |
| |
| rx_len = (xfer->len / bytes_per_word); |
| } |
| |
| if (xfer->tx_buf && xfer->rx_buf) { |
| cmd = SPI_FULL_DUPLEX; |
| tx_nent += 2; |
| rx_nent++; |
| } else if (xfer->tx_buf) { |
| cmd = SPI_TX_ONLY; |
| tx_nent += 2; |
| rx_len = 0; |
| } else if (xfer->rx_buf) { |
| cmd = SPI_RX_ONLY; |
| tx_nent++; |
| rx_nent++; |
| } |
| |
| cs |= spi_slv->chip_select; |
| if (!list_is_last(&xfer->transfer_list, &spi->cur_msg->transfers)) |
| go_flags |= FRAGMENTATION; |
| go_tre = setup_go_tre(cmd, cs, rx_len, go_flags, mas); |
| |
| sg_init_table(xfer_tx_sg, tx_nent); |
| if (rx_nent) |
| sg_init_table(xfer_rx_sg, rx_nent); |
| |
| if (c0_tre) |
| sg_set_buf(xfer_tx_sg++, c0_tre, sizeof(*c0_tre)); |
| |
| sg_set_buf(xfer_tx_sg++, go_tre, sizeof(*go_tre)); |
| mas->gsi[mas->num_xfers].desc_cb.spi = spi; |
| mas->gsi[mas->num_xfers].desc_cb.xfer = xfer; |
| if (cmd & SPI_RX_ONLY) { |
| rx_tre = &mas->gsi[mas->num_xfers].rx_dma_tre; |
| rx_tre = setup_dma_tre(rx_tre, xfer->rx_dma, xfer->len, mas, 0); |
| if (IS_ERR_OR_NULL(rx_tre)) { |
| dev_err(mas->dev, "Err setting up rx tre\n"); |
| return PTR_ERR(rx_tre); |
| } |
| sg_set_buf(xfer_rx_sg, rx_tre, sizeof(*rx_tre)); |
| mas->gsi[mas->num_xfers].rx_desc = |
| dmaengine_prep_slave_sg(mas->rx, |
| &mas->gsi[mas->num_xfers].rx_sg, rx_nent, |
| DMA_DEV_TO_MEM, flags); |
| if (IS_ERR_OR_NULL(mas->gsi[mas->num_xfers].rx_desc)) { |
| dev_err(mas->dev, "Err setting up rx desc\n"); |
| return -EIO; |
| } |
| mas->gsi[mas->num_xfers].rx_desc->callback = |
| spi_gsi_rx_callback; |
| mas->gsi[mas->num_xfers].rx_desc->callback_param = |
| &mas->gsi[mas->num_xfers].rx_cb_param; |
| mas->gsi[mas->num_xfers].rx_cb_param.userdata = |
| &mas->gsi[mas->num_xfers].desc_cb; |
| mas->num_rx_eot++; |
| } |
| |
| if (cmd & SPI_TX_ONLY) { |
| tx_tre = &mas->gsi[mas->num_xfers].tx_dma_tre; |
| tx_tre = setup_dma_tre(tx_tre, xfer->tx_dma, xfer->len, mas, 1); |
| if (IS_ERR_OR_NULL(tx_tre)) { |
| dev_err(mas->dev, "Err setting up tx tre\n"); |
| return PTR_ERR(tx_tre); |
| } |
| sg_set_buf(xfer_tx_sg++, tx_tre, sizeof(*tx_tre)); |
| mas->num_tx_eot++; |
| } |
| mas->gsi[mas->num_xfers].tx_desc = dmaengine_prep_slave_sg(mas->tx, |
| mas->gsi[mas->num_xfers].tx_sg, tx_nent, |
| DMA_MEM_TO_DEV, flags); |
| if (IS_ERR_OR_NULL(mas->gsi[mas->num_xfers].tx_desc)) { |
| dev_err(mas->dev, "Err setting up tx desc\n"); |
| return -EIO; |
| } |
| mas->gsi[mas->num_xfers].tx_desc->callback = spi_gsi_tx_callback; |
| mas->gsi[mas->num_xfers].tx_desc->callback_param = |
| &mas->gsi[mas->num_xfers].tx_cb_param; |
| mas->gsi[mas->num_xfers].tx_cb_param.userdata = |
| &mas->gsi[mas->num_xfers].desc_cb; |
| mas->gsi[mas->num_xfers].tx_cookie = |
| dmaengine_submit(mas->gsi[mas->num_xfers].tx_desc); |
| if (mas->num_rx_eot) |
| mas->gsi[mas->num_xfers].rx_cookie = |
| dmaengine_submit(mas->gsi[mas->num_xfers].rx_desc); |
| dma_async_issue_pending(mas->tx); |
| if (mas->num_rx_eot) |
| dma_async_issue_pending(mas->rx); |
| mas->num_xfers++; |
| return ret; |
| } |
| |
| static int spi_geni_map_buf(struct spi_geni_master *mas, |
| struct spi_message *msg) |
| { |
| struct spi_transfer *xfer; |
| int ret = 0; |
| |
| list_for_each_entry(xfer, &msg->transfers, transfer_list) { |
| if (xfer->rx_buf) { |
| ret = geni_se_iommu_map_buf(mas->wrapper_dev, |
| &xfer->rx_dma, xfer->rx_buf, |
| xfer->len, DMA_FROM_DEVICE); |
| if (ret) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Mapping Rx buffer %d\n", __func__, ret); |
| return ret; |
| } |
| } |
| |
| if (xfer->tx_buf) { |
| ret = geni_se_iommu_map_buf(mas->wrapper_dev, |
| &xfer->tx_dma, |
| (void *)xfer->tx_buf, |
| xfer->len, DMA_TO_DEVICE); |
| if (ret) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "%s: Mapping Tx buffer %d\n", __func__, ret); |
| return ret; |
| } |
| } |
| }; |
| return 0; |
| } |
| |
| static void spi_geni_unmap_buf(struct spi_geni_master *mas, |
| struct spi_message *msg) |
| { |
| struct spi_transfer *xfer; |
| |
| list_for_each_entry(xfer, &msg->transfers, transfer_list) { |
| if (xfer->rx_buf) |
| geni_se_iommu_unmap_buf(mas->wrapper_dev, &xfer->rx_dma, |
| xfer->len, DMA_FROM_DEVICE); |
| if (xfer->tx_buf) |
| geni_se_iommu_unmap_buf(mas->wrapper_dev, &xfer->tx_dma, |
| xfer->len, DMA_TO_DEVICE); |
| }; |
| } |
| |
| static int spi_geni_prepare_message(struct spi_master *spi, |
| struct spi_message *spi_msg) |
| { |
| int ret = 0; |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| |
| mas->cur_xfer_mode = select_xfer_mode(spi, spi_msg); |
| |
| if (mas->cur_xfer_mode == FIFO_MODE) { |
| geni_se_select_mode(mas->base, FIFO_MODE); |
| reinit_completion(&mas->xfer_done); |
| setup_fifo_params(spi_msg->spi, spi); |
| } else if (mas->cur_xfer_mode == GSI_DMA) { |
| mas->num_tx_eot = 0; |
| mas->num_rx_eot = 0; |
| mas->num_xfers = 0; |
| reinit_completion(&mas->tx_cb); |
| reinit_completion(&mas->rx_cb); |
| memset(mas->gsi, 0, |
| (sizeof(struct spi_geni_gsi) * NUM_SPI_XFER)); |
| geni_se_select_mode(mas->base, GSI_DMA); |
| dmaengine_resume(mas->tx); |
| ret = spi_geni_map_buf(mas, spi_msg); |
| } else { |
| dev_err(mas->dev, "%s: Couldn't select mode %d", __func__, |
| mas->cur_xfer_mode); |
| ret = -EINVAL; |
| } |
| return ret; |
| } |
| |
| static int spi_geni_unprepare_message(struct spi_master *spi_mas, |
| struct spi_message *spi_msg) |
| { |
| struct spi_geni_master *mas = spi_master_get_devdata(spi_mas); |
| |
| mas->cur_speed_hz = 0; |
| mas->cur_word_len = 0; |
| if (mas->cur_xfer_mode == GSI_DMA) { |
| dmaengine_pause(mas->tx); |
| spi_geni_unmap_buf(mas, spi_msg); |
| } |
| return 0; |
| } |
| |
| static int spi_geni_prepare_transfer_hardware(struct spi_master *spi) |
| { |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| int ret = 0; |
| u32 max_speed = spi->cur_msg->spi->max_speed_hz; |
| struct se_geni_rsc *rsc = &mas->spi_rsc; |
| |
| /* Adjust the AB/IB based on the max speed of the slave.*/ |
| rsc->ib = max_speed * DEFAULT_BUS_WIDTH; |
| rsc->ab = max_speed * DEFAULT_BUS_WIDTH; |
| ret = pm_runtime_get_sync(mas->dev); |
| if (ret < 0) { |
| dev_err(mas->dev, "Error enabling SE resources\n"); |
| pm_runtime_put_noidle(mas->dev); |
| goto exit_prepare_transfer_hardware; |
| } else { |
| ret = 0; |
| } |
| |
| if (unlikely(!mas->setup)) { |
| int proto = get_se_proto(mas->base); |
| unsigned int major; |
| unsigned int minor; |
| unsigned int step; |
| int hw_ver; |
| |
| if (unlikely(proto != SPI)) { |
| dev_err(mas->dev, "Invalid proto %d\n", proto); |
| return -ENXIO; |
| } |
| geni_se_init(mas->base, 0x0, (mas->tx_fifo_depth - 2)); |
| mas->tx_fifo_depth = get_tx_fifo_depth(mas->base); |
| mas->rx_fifo_depth = get_rx_fifo_depth(mas->base); |
| mas->tx_fifo_width = get_tx_fifo_width(mas->base); |
| mas->oversampling = 1; |
| /* Transmit an entire FIFO worth of data per IRQ */ |
| mas->tx_wm = 1; |
| |
| mas->tx = dma_request_slave_channel(mas->dev, "tx"); |
| if (IS_ERR_OR_NULL(mas->tx)) { |
| dev_info(mas->dev, "Failed to get tx DMA ch %ld", |
| PTR_ERR(mas->tx)); |
| } else { |
| mas->rx = dma_request_slave_channel(mas->dev, "rx"); |
| if (IS_ERR_OR_NULL(mas->rx)) { |
| dev_info(mas->dev, "Failed to get rx DMA ch %ld", |
| PTR_ERR(mas->rx)); |
| dma_release_channel(mas->tx); |
| } |
| mas->gsi = devm_kzalloc(mas->dev, |
| (sizeof(struct spi_geni_gsi) * NUM_SPI_XFER), |
| GFP_KERNEL); |
| if (IS_ERR_OR_NULL(mas->gsi)) { |
| dev_err(mas->dev, "Failed to get GSI mem\n"); |
| dma_release_channel(mas->tx); |
| dma_release_channel(mas->rx); |
| mas->tx = NULL; |
| mas->rx = NULL; |
| goto setup_ipc; |
| } |
| mas->tx_event.init.callback = spi_gsi_ch_cb; |
| mas->tx_event.init.cb_param = spi; |
| mas->tx_event.cmd = MSM_GPI_INIT; |
| mas->tx->private = &mas->tx_event; |
| mas->rx_event.init.callback = spi_gsi_ch_cb; |
| mas->rx_event.init.cb_param = spi; |
| mas->rx_event.cmd = MSM_GPI_INIT; |
| mas->rx->private = &mas->rx_event; |
| if (dmaengine_slave_config(mas->tx, NULL)) { |
| dev_err(mas->dev, "Failed to Config Tx\n"); |
| dma_release_channel(mas->tx); |
| dma_release_channel(mas->rx); |
| mas->tx = NULL; |
| mas->rx = NULL; |
| goto setup_ipc; |
| } |
| if (dmaengine_slave_config(mas->rx, NULL)) { |
| dev_err(mas->dev, "Failed to Config Rx\n"); |
| dma_release_channel(mas->tx); |
| dma_release_channel(mas->rx); |
| mas->tx = NULL; |
| mas->rx = NULL; |
| goto setup_ipc; |
| } |
| |
| } |
| setup_ipc: |
| mas->ipc = ipc_log_context_create(4, dev_name(mas->dev), 0); |
| dev_info(mas->dev, "tx_fifo %d rx_fifo %d tx_width %d\n", |
| mas->tx_fifo_depth, mas->rx_fifo_depth, |
| mas->tx_fifo_width); |
| mas->setup = true; |
| hw_ver = geni_se_qupv3_hw_version(mas->wrapper_dev, &major, |
| &minor, &step); |
| if (hw_ver) |
| dev_err(mas->dev, "%s:Err getting HW version %d\n", |
| __func__, hw_ver); |
| else { |
| if ((major == 1) && (minor == 0)) |
| mas->oversampling = 2; |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:Major:%d Minor:%d step:%dos%d\n", |
| __func__, major, minor, step, mas->oversampling); |
| } |
| } |
| exit_prepare_transfer_hardware: |
| return ret; |
| } |
| |
| static int spi_geni_unprepare_transfer_hardware(struct spi_master *spi) |
| { |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| |
| pm_runtime_put_sync(mas->dev); |
| return 0; |
| } |
| |
| static void setup_fifo_xfer(struct spi_transfer *xfer, |
| struct spi_geni_master *mas, u16 mode, |
| struct spi_master *spi) |
| { |
| u32 m_cmd = 0; |
| u32 m_param = 0; |
| u32 spi_tx_cfg = geni_read_reg(mas->base, SE_SPI_TRANS_CFG); |
| u32 trans_len = 0; |
| |
| if (xfer->bits_per_word != mas->cur_word_len) { |
| spi_setup_word_len(mas, mode, xfer->bits_per_word); |
| mas->cur_word_len = xfer->bits_per_word; |
| } |
| |
| /* Speed and bits per word can be overridden per transfer */ |
| if (xfer->speed_hz != mas->cur_speed_hz) { |
| int ret = 0; |
| u32 clk_sel = geni_read_reg(mas->base, SE_GENI_CLK_SEL); |
| u32 m_clk_cfg = geni_read_reg(mas->base, GENI_SER_M_CLK_CFG); |
| int idx = 0; |
| int div = 0; |
| |
| ret = get_spi_clk_cfg(xfer->speed_hz, mas, &idx, &div); |
| if (ret) { |
| dev_err(mas->dev, "%s:Err setting clks:%d\n", |
| __func__, ret); |
| return; |
| } |
| mas->cur_speed_hz = xfer->speed_hz; |
| clk_sel |= (idx & CLK_SEL_MSK); |
| m_clk_cfg |= ((div << CLK_DIV_SHFT) | SER_CLK_EN); |
| geni_write_reg(clk_sel, mas->base, SE_GENI_CLK_SEL); |
| geni_write_reg(m_clk_cfg, mas->base, GENI_SER_M_CLK_CFG); |
| } |
| |
| mas->tx_rem_bytes = 0; |
| mas->rx_rem_bytes = 0; |
| if (xfer->tx_buf && xfer->rx_buf) |
| m_cmd = SPI_FULL_DUPLEX; |
| else if (xfer->tx_buf) |
| m_cmd = SPI_TX_ONLY; |
| else if (xfer->rx_buf) |
| m_cmd = SPI_RX_ONLY; |
| |
| spi_tx_cfg &= ~CS_TOGGLE; |
| if (xfer->cs_change) |
| spi_tx_cfg |= CS_TOGGLE; |
| if (!(mas->cur_word_len % MIN_WORD_LEN)) { |
| trans_len = |
| ((xfer->len << 3) / mas->cur_word_len) & TRANS_LEN_MSK; |
| } else { |
| int bytes_per_word = (mas->cur_word_len / BITS_PER_BYTE) + 1; |
| |
| trans_len = (xfer->len / bytes_per_word) & TRANS_LEN_MSK; |
| } |
| if (!list_is_last(&xfer->transfer_list, &spi->cur_msg->transfers)) |
| m_param |= FRAGMENTATION; |
| |
| mas->cur_xfer = xfer; |
| if (m_cmd & SPI_TX_ONLY) { |
| mas->tx_rem_bytes = xfer->len; |
| geni_write_reg(trans_len, mas->base, SE_SPI_TX_TRANS_LEN); |
| } |
| |
| if (m_cmd & SPI_RX_ONLY) { |
| geni_write_reg(trans_len, mas->base, SE_SPI_RX_TRANS_LEN); |
| mas->rx_rem_bytes = xfer->len; |
| } |
| geni_write_reg(spi_tx_cfg, mas->base, SE_SPI_TRANS_CFG); |
| geni_setup_m_cmd(mas->base, m_cmd, m_param); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: trans_len %d xferlen%d tx_cfg 0x%x cmd 0x%x\n", |
| __func__, trans_len, xfer->len, spi_tx_cfg, m_cmd); |
| if (m_cmd & SPI_TX_ONLY) |
| geni_write_reg(mas->tx_wm, mas->base, SE_GENI_TX_WATERMARK_REG); |
| /* Ensure all writes are done before the WM interrupt */ |
| mb(); |
| } |
| |
| static void handle_fifo_timeout(struct spi_geni_master *mas) |
| { |
| unsigned long timeout; |
| |
| geni_se_dump_dbg_regs(&mas->spi_rsc, mas->base, mas->ipc); |
| reinit_completion(&mas->xfer_done); |
| geni_cancel_m_cmd(mas->base); |
| geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG); |
| /* Ensure cmd cancel is written */ |
| mb(); |
| timeout = wait_for_completion_timeout(&mas->xfer_done, HZ); |
| if (!timeout) { |
| reinit_completion(&mas->xfer_done); |
| geni_abort_m_cmd(mas->base); |
| /* Ensure cmd abort is written */ |
| mb(); |
| timeout = wait_for_completion_timeout(&mas->xfer_done, |
| HZ); |
| if (!timeout) |
| dev_err(mas->dev, |
| "Failed to cancel/abort m_cmd\n"); |
| } |
| } |
| |
| static int spi_geni_transfer_one(struct spi_master *spi, |
| struct spi_device *slv, |
| struct spi_transfer *xfer) |
| { |
| int ret = 0; |
| struct spi_geni_master *mas = spi_master_get_devdata(spi); |
| unsigned long timeout; |
| |
| if ((xfer->tx_buf == NULL) && (xfer->rx_buf == NULL)) { |
| dev_err(mas->dev, "Invalid xfer both tx rx are NULL\n"); |
| return -EINVAL; |
| } |
| |
| if (mas->cur_xfer_mode == FIFO_MODE) { |
| setup_fifo_xfer(xfer, mas, slv->mode, spi); |
| timeout = wait_for_completion_timeout(&mas->xfer_done, |
| msecs_to_jiffies(SPI_XFER_TIMEOUT_MS)); |
| if (!timeout) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "Xfer[len %d tx %pK rx %pK n %d] timed out.\n", |
| xfer->len, xfer->tx_buf, |
| xfer->rx_buf, |
| xfer->bits_per_word); |
| mas->cur_xfer = NULL; |
| ret = -ETIMEDOUT; |
| goto err_fifo_geni_transfer_one; |
| } |
| } else { |
| setup_gsi_xfer(xfer, mas, slv, spi); |
| if ((mas->num_xfers >= NUM_SPI_XFER) || |
| (list_is_last(&xfer->transfer_list, |
| &spi->cur_msg->transfers))) { |
| int i; |
| |
| for (i = 0 ; i < mas->num_tx_eot; i++) { |
| timeout = |
| wait_for_completion_interruptible_timeout( |
| &mas->tx_cb, |
| msecs_to_jiffies(SPI_XFER_TIMEOUT_MS)); |
| if (timeout <= 0) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "Tx[%d] timeout%lu\n", i, timeout); |
| ret = -ETIMEDOUT; |
| goto err_gsi_geni_transfer_one; |
| } |
| } |
| for (i = 0 ; i < mas->num_rx_eot; i++) { |
| timeout = |
| wait_for_completion_interruptible_timeout( |
| &mas->rx_cb, |
| msecs_to_jiffies(SPI_XFER_TIMEOUT_MS)); |
| if (timeout <= 0) { |
| GENI_SE_ERR(mas->ipc, true, mas->dev, |
| "Rx[%d] timeout%lu\n", i, timeout); |
| ret = -ETIMEDOUT; |
| goto err_gsi_geni_transfer_one; |
| } |
| } |
| if (mas->qn_err) { |
| ret = -EIO; |
| mas->qn_err = false; |
| goto err_gsi_geni_transfer_one; |
| } |
| } |
| } |
| return ret; |
| err_gsi_geni_transfer_one: |
| geni_se_dump_dbg_regs(&mas->spi_rsc, mas->base, mas->ipc); |
| dmaengine_terminate_all(mas->tx); |
| return ret; |
| err_fifo_geni_transfer_one: |
| handle_fifo_timeout(mas); |
| return ret; |
| } |
| |
| static void geni_spi_handle_tx(struct spi_geni_master *mas) |
| { |
| int i = 0; |
| int tx_fifo_width = (mas->tx_fifo_width >> 3); |
| int max_bytes = 0; |
| const u8 *tx_buf = NULL; |
| |
| if (!mas->cur_xfer) |
| return; |
| |
| /* |
| * For non-byte aligned bits-per-word values: |
| * Assumption is that each SPI word will be accomodated in |
| * ceil (bits_per_word / bits_per_byte) |
| * and the next SPI word starts at the next byte. |
| * In such cases, we can fit 1 SPI word per FIFO word so adjust the |
| * max byte that can be sent per IRQ accordingly. |
| */ |
| if ((mas->tx_fifo_width % mas->cur_word_len)) |
| max_bytes = (mas->tx_fifo_depth - mas->tx_wm) * |
| ((mas->cur_word_len / BITS_PER_BYTE) + 1); |
| else |
| max_bytes = (mas->tx_fifo_depth - mas->tx_wm) * tx_fifo_width; |
| tx_buf = mas->cur_xfer->tx_buf; |
| tx_buf += (mas->cur_xfer->len - mas->tx_rem_bytes); |
| max_bytes = min_t(int, mas->tx_rem_bytes, max_bytes); |
| while (i < max_bytes) { |
| int j; |
| u32 fifo_word = 0; |
| u8 *fifo_byte; |
| int bytes_per_fifo = tx_fifo_width; |
| int bytes_to_write = 0; |
| |
| if ((mas->tx_fifo_width % mas->cur_word_len)) |
| bytes_per_fifo = |
| (mas->cur_word_len / BITS_PER_BYTE) + 1; |
| bytes_to_write = min_t(int, (max_bytes - i), bytes_per_fifo); |
| fifo_byte = (u8 *)&fifo_word; |
| for (j = 0; j < bytes_to_write; j++) |
| fifo_byte[j] = tx_buf[i++]; |
| geni_write_reg(fifo_word, mas->base, SE_GENI_TX_FIFOn); |
| /* Ensure FIFO writes are written in order */ |
| mb(); |
| } |
| mas->tx_rem_bytes -= max_bytes; |
| if (!mas->tx_rem_bytes) { |
| geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG); |
| /* Barrier here before return to prevent further ISRs */ |
| mb(); |
| } |
| } |
| |
| static void geni_spi_handle_rx(struct spi_geni_master *mas) |
| { |
| int i = 0; |
| int fifo_width = (mas->tx_fifo_width >> 3); |
| u32 rx_fifo_status = geni_read_reg(mas->base, SE_GENI_RX_FIFO_STATUS); |
| int rx_bytes = 0; |
| int rx_wc = 0; |
| u8 *rx_buf = NULL; |
| |
| if (!mas->cur_xfer) |
| return; |
| |
| rx_buf = mas->cur_xfer->rx_buf; |
| rx_wc = (rx_fifo_status & RX_FIFO_WC_MSK); |
| if (rx_fifo_status & RX_LAST) { |
| int rx_last_byte_valid = |
| (rx_fifo_status & RX_LAST_BYTE_VALID_MSK) |
| >> RX_LAST_BYTE_VALID_SHFT; |
| if (rx_last_byte_valid && (rx_last_byte_valid < 4)) { |
| rx_wc -= 1; |
| rx_bytes += rx_last_byte_valid; |
| } |
| } |
| if (!(mas->tx_fifo_width % mas->cur_word_len)) |
| rx_bytes += rx_wc * fifo_width; |
| else |
| rx_bytes += rx_wc * |
| ((mas->cur_word_len / BITS_PER_BYTE) + 1); |
| rx_bytes = min_t(int, mas->rx_rem_bytes, rx_bytes); |
| rx_buf += (mas->cur_xfer->len - mas->rx_rem_bytes); |
| while (i < rx_bytes) { |
| u32 fifo_word = 0; |
| u8 *fifo_byte; |
| int bytes_per_fifo = fifo_width; |
| int read_bytes = 0; |
| int j; |
| |
| if ((mas->tx_fifo_width % mas->cur_word_len)) |
| bytes_per_fifo = |
| (mas->cur_word_len / BITS_PER_BYTE) + 1; |
| read_bytes = min_t(int, (rx_bytes - i), bytes_per_fifo); |
| fifo_word = geni_read_reg(mas->base, SE_GENI_RX_FIFOn); |
| fifo_byte = (u8 *)&fifo_word; |
| for (j = 0; j < read_bytes; j++) |
| rx_buf[i++] = fifo_byte[j]; |
| } |
| mas->rx_rem_bytes -= rx_bytes; |
| } |
| |
| static irqreturn_t geni_spi_irq(int irq, void *dev) |
| { |
| struct spi_geni_master *mas = dev; |
| u32 m_irq = 0; |
| |
| if (pm_runtime_status_suspended(dev)) { |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s: device is suspended\n", __func__); |
| goto exit_geni_spi_irq; |
| } |
| m_irq = geni_read_reg(mas->base, SE_GENI_M_IRQ_STATUS); |
| if ((m_irq & M_RX_FIFO_WATERMARK_EN) || (m_irq & M_RX_FIFO_LAST_EN)) |
| geni_spi_handle_rx(mas); |
| |
| if ((m_irq & M_TX_FIFO_WATERMARK_EN)) |
| geni_spi_handle_tx(mas); |
| |
| if ((m_irq & M_CMD_DONE_EN) || (m_irq & M_CMD_CANCEL_EN) || |
| (m_irq & M_CMD_ABORT_EN)) { |
| complete(&mas->xfer_done); |
| /* |
| * If this happens, then a CMD_DONE came before all the buffer |
| * bytes were sent out. This is unusual, log this condition and |
| * disable the WM interrupt to prevent the system from stalling |
| * due an interrupt storm. |
| * If this happens when all Rx bytes haven't been received, log |
| * the condition. |
| */ |
| if (mas->tx_rem_bytes) { |
| geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG); |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:Premature Done.tx_rem%d bpw%d\n", |
| __func__, mas->tx_rem_bytes, mas->cur_word_len); |
| } |
| if (mas->rx_rem_bytes) |
| GENI_SE_DBG(mas->ipc, false, mas->dev, |
| "%s:Premature Done.rx_rem%d bpw%d\n", |
| __func__, mas->rx_rem_bytes, mas->cur_word_len); |
| } |
| exit_geni_spi_irq: |
| geni_write_reg(m_irq, mas->base, SE_GENI_M_IRQ_CLEAR); |
| return IRQ_HANDLED; |
| } |
| |
| static int spi_geni_probe(struct platform_device *pdev) |
| { |
| int ret; |
| struct spi_master *spi; |
| struct spi_geni_master *geni_mas; |
| struct se_geni_rsc *rsc; |
| struct resource *res; |
| struct platform_device *wrapper_pdev; |
| struct device_node *wrapper_ph_node; |
| |
| spi = spi_alloc_master(&pdev->dev, sizeof(struct spi_geni_master)); |
| if (!spi) { |
| ret = -ENOMEM; |
| dev_err(&pdev->dev, "Failed to alloc spi struct\n"); |
| goto spi_geni_probe_err; |
| } |
| |
| platform_set_drvdata(pdev, spi); |
| geni_mas = spi_master_get_devdata(spi); |
| rsc = &geni_mas->spi_rsc; |
| geni_mas->dev = &pdev->dev; |
| spi->dev.of_node = pdev->dev.of_node; |
| wrapper_ph_node = of_parse_phandle(pdev->dev.of_node, |
| "qcom,wrapper-core", 0); |
| if (IS_ERR_OR_NULL(wrapper_ph_node)) { |
| ret = PTR_ERR(wrapper_ph_node); |
| dev_err(&pdev->dev, "No wrapper core defined\n"); |
| goto spi_geni_probe_err; |
| } |
| wrapper_pdev = of_find_device_by_node(wrapper_ph_node); |
| of_node_put(wrapper_ph_node); |
| if (IS_ERR_OR_NULL(wrapper_pdev)) { |
| ret = PTR_ERR(wrapper_pdev); |
| dev_err(&pdev->dev, "Cannot retrieve wrapper device\n"); |
| goto spi_geni_probe_err; |
| } |
| geni_mas->wrapper_dev = &wrapper_pdev->dev; |
| geni_mas->spi_rsc.wrapper_dev = &wrapper_pdev->dev; |
| ret = geni_se_resources_init(rsc, SPI_CORE2X_VOTE, |
| (DEFAULT_SE_CLK * DEFAULT_BUS_WIDTH)); |
| if (ret) { |
| dev_err(&pdev->dev, "Error geni_se_resources_init\n"); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->geni_pinctrl = devm_pinctrl_get(&pdev->dev); |
| if (IS_ERR_OR_NULL(rsc->geni_pinctrl)) { |
| dev_err(&pdev->dev, "No pinctrl config specified!\n"); |
| ret = PTR_ERR(rsc->geni_pinctrl); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->geni_gpio_active = pinctrl_lookup_state(rsc->geni_pinctrl, |
| PINCTRL_DEFAULT); |
| if (IS_ERR_OR_NULL(rsc->geni_gpio_active)) { |
| dev_err(&pdev->dev, "No default config specified!\n"); |
| ret = PTR_ERR(rsc->geni_gpio_active); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->geni_gpio_sleep = pinctrl_lookup_state(rsc->geni_pinctrl, |
| PINCTRL_SLEEP); |
| if (IS_ERR_OR_NULL(rsc->geni_gpio_sleep)) { |
| dev_err(&pdev->dev, "No sleep config specified!\n"); |
| ret = PTR_ERR(rsc->geni_gpio_sleep); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->se_clk = devm_clk_get(&pdev->dev, "se-clk"); |
| if (IS_ERR(rsc->se_clk)) { |
| ret = PTR_ERR(rsc->se_clk); |
| dev_err(&pdev->dev, "Err getting SE Core clk %d\n", ret); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->m_ahb_clk = devm_clk_get(&pdev->dev, "m-ahb"); |
| if (IS_ERR(rsc->m_ahb_clk)) { |
| ret = PTR_ERR(rsc->m_ahb_clk); |
| dev_err(&pdev->dev, "Err getting M AHB clk %d\n", ret); |
| goto spi_geni_probe_err; |
| } |
| |
| rsc->s_ahb_clk = devm_clk_get(&pdev->dev, "s-ahb"); |
| if (IS_ERR(rsc->s_ahb_clk)) { |
| ret = PTR_ERR(rsc->s_ahb_clk); |
| dev_err(&pdev->dev, "Err getting S AHB clk %d\n", ret); |
| goto spi_geni_probe_err; |
| } |
| |
| if (of_property_read_u32(pdev->dev.of_node, "spi-max-frequency", |
| &spi->max_speed_hz)) { |
| dev_err(&pdev->dev, "Max frequency not specified.\n"); |
| ret = -ENXIO; |
| goto spi_geni_probe_err; |
| } |
| |
| res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "se_phys"); |
| if (!res) { |
| ret = -ENXIO; |
| dev_err(&pdev->dev, "Err getting IO region\n"); |
| goto spi_geni_probe_err; |
| } |
| |
| geni_mas->phys_addr = res->start; |
| geni_mas->size = resource_size(res); |
| geni_mas->base = devm_ioremap(&pdev->dev, res->start, |
| resource_size(res)); |
| if (!geni_mas->base) { |
| ret = -ENOMEM; |
| dev_err(&pdev->dev, "Err IO mapping iomem\n"); |
| goto spi_geni_probe_err; |
| } |
| |
| geni_mas->irq = platform_get_irq(pdev, 0); |
| if (geni_mas->irq < 0) { |
| dev_err(&pdev->dev, "Err getting IRQ\n"); |
| ret = geni_mas->irq; |
| goto spi_geni_probe_unmap; |
| } |
| ret = devm_request_irq(&pdev->dev, geni_mas->irq, geni_spi_irq, |
| IRQF_TRIGGER_HIGH, "spi_geni", geni_mas); |
| if (ret) { |
| dev_err(&pdev->dev, "Request_irq failed:%d: err:%d\n", |
| geni_mas->irq, ret); |
| goto spi_geni_probe_unmap; |
| } |
| |
| spi->mode_bits = (SPI_CPOL | SPI_CPHA | SPI_LOOP | SPI_CS_HIGH); |
| spi->bits_per_word_mask = SPI_BPW_RANGE_MASK(4, 32); |
| spi->num_chipselect = SPI_NUM_CHIPSELECT; |
| spi->prepare_transfer_hardware = spi_geni_prepare_transfer_hardware; |
| spi->prepare_message = spi_geni_prepare_message; |
| spi->unprepare_message = spi_geni_unprepare_message; |
| spi->transfer_one = spi_geni_transfer_one; |
| spi->unprepare_transfer_hardware |
| = spi_geni_unprepare_transfer_hardware; |
| spi->auto_runtime_pm = false; |
| |
| init_completion(&geni_mas->xfer_done); |
| init_completion(&geni_mas->tx_cb); |
| init_completion(&geni_mas->rx_cb); |
| pm_runtime_enable(&pdev->dev); |
| ret = spi_register_master(spi); |
| if (ret) { |
| dev_err(&pdev->dev, "Failed to register SPI master\n"); |
| goto spi_geni_probe_unmap; |
| } |
| return ret; |
| spi_geni_probe_unmap: |
| devm_iounmap(&pdev->dev, geni_mas->base); |
| spi_geni_probe_err: |
| spi_master_put(spi); |
| return ret; |
| } |
| |
| static int spi_geni_remove(struct platform_device *pdev) |
| { |
| struct spi_master *master = platform_get_drvdata(pdev); |
| struct spi_geni_master *geni_mas = spi_master_get_devdata(master); |
| |
| spi_unregister_master(master); |
| se_geni_resources_off(&geni_mas->spi_rsc); |
| pm_runtime_put_noidle(&pdev->dev); |
| pm_runtime_disable(&pdev->dev); |
| return 0; |
| } |
| |
| #ifdef CONFIG_PM |
| static int spi_geni_runtime_suspend(struct device *dev) |
| { |
| int ret = 0; |
| struct spi_master *spi = get_spi_master(dev); |
| struct spi_geni_master *geni_mas = spi_master_get_devdata(spi); |
| |
| ret = se_geni_resources_off(&geni_mas->spi_rsc); |
| return ret; |
| } |
| |
| static int spi_geni_runtime_resume(struct device *dev) |
| { |
| int ret = 0; |
| struct spi_master *spi = get_spi_master(dev); |
| struct spi_geni_master *geni_mas = spi_master_get_devdata(spi); |
| |
| ret = se_geni_resources_on(&geni_mas->spi_rsc); |
| return ret; |
| } |
| |
| static int spi_geni_resume(struct device *dev) |
| { |
| return 0; |
| } |
| |
| static int spi_geni_suspend(struct device *dev) |
| { |
| if (!pm_runtime_status_suspended(dev)) |
| return -EBUSY; |
| return 0; |
| } |
| #else |
| static int spi_geni_runtime_suspend(struct device *dev) |
| { |
| return 0; |
| } |
| |
| static int spi_geni_runtime_resume(struct device *dev) |
| { |
| return 0; |
| } |
| |
| static int spi_geni_resume(struct device *dev) |
| { |
| return 0; |
| } |
| |
| static int spi_geni_suspend(struct device *dev) |
| { |
| return 0; |
| } |
| #endif |
| |
| static const struct dev_pm_ops spi_geni_pm_ops = { |
| SET_RUNTIME_PM_OPS(spi_geni_runtime_suspend, |
| spi_geni_runtime_resume, NULL) |
| SET_SYSTEM_SLEEP_PM_OPS(spi_geni_suspend, spi_geni_resume) |
| }; |
| |
| static const struct of_device_id spi_geni_dt_match[] = { |
| { .compatible = "qcom,spi-geni" }, |
| {} |
| }; |
| |
| static struct platform_driver spi_geni_driver = { |
| .probe = spi_geni_probe, |
| .remove = spi_geni_remove, |
| .driver = { |
| .name = "spi_geni", |
| .pm = &spi_geni_pm_ops, |
| .of_match_table = spi_geni_dt_match, |
| }, |
| }; |
| module_platform_driver(spi_geni_driver); |
| |
| MODULE_LICENSE("GPL v2"); |
| MODULE_ALIAS("platform:spi_geni"); |