blob: de3dc6970c70e39c30bb334de7b3f51096c594cd [file] [log] [blame]
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001/*
2 * Copyright (c) 2017, The Linux Foundation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 and
6 * only version 2 as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope that it will be useful,
9 * but WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 * GNU General Public License for more details.
12 *
13 */
14#include <linux/clk.h>
Girish Mahadevan9a7c9442017-08-15 12:10:09 -060015#include <linux/dmaengine.h>
16#include <linux/dma-mapping.h>
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070017#include <linux/interrupt.h>
Girish Mahadevan9a7c9442017-08-15 12:10:09 -060018#include <linux/ipc_logging.h>
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070019#include <linux/io.h>
20#include <linux/module.h>
21#include <linux/of.h>
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -060022#include <linux/of_platform.h>
Karthikeyan Ramasubramanian9a633402017-04-06 16:01:11 -060023#include <linux/pm_runtime.h>
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070024#include <linux/qcom-geni-se.h>
Girish Mahadevan9a7c9442017-08-15 12:10:09 -060025#include <linux/msm_gpi.h>
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070026#include <linux/spi/spi.h>
27
28#define SPI_NUM_CHIPSELECT (4)
29#define SPI_XFER_TIMEOUT_MS (250)
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070030/* SPI SE specific registers */
31#define SE_SPI_CPHA (0x224)
32#define SE_SPI_LOOPBACK (0x22C)
33#define SE_SPI_CPOL (0x230)
34#define SE_SPI_DEMUX_OUTPUT_INV (0x24C)
35#define SE_SPI_DEMUX_SEL (0x250)
36#define SE_SPI_TRANS_CFG (0x25C)
37#define SE_SPI_WORD_LEN (0x268)
38#define SE_SPI_TX_TRANS_LEN (0x26C)
39#define SE_SPI_RX_TRANS_LEN (0x270)
40#define SE_SPI_PRE_POST_CMD_DLY (0x274)
41#define SE_SPI_DELAY_COUNTERS (0x278)
42
43/* SE_SPI_CPHA register fields */
44#define CPHA (BIT(0))
45
46/* SE_SPI_LOOPBACK register fields */
47#define LOOPBACK_ENABLE (0x1)
48#define NORMAL_MODE (0x0)
49#define LOOPBACK_MSK (GENMASK(1, 0))
50
51/* SE_SPI_CPOL register fields */
52#define CPOL (BIT(2))
53
54/* SE_SPI_DEMUX_OUTPUT_INV register fields */
55#define CS_DEMUX_OUTPUT_INV_MSK (GENMASK(3, 0))
56
57/* SE_SPI_DEMUX_SEL register fields */
58#define CS_DEMUX_OUTPUT_SEL (GENMASK(3, 0))
59
60/* SE_SPI_TX_TRANS_CFG register fields */
61#define CS_TOGGLE (BIT(0))
62
63/* SE_SPI_WORD_LEN register fields */
64#define WORD_LEN_MSK (GENMASK(9, 0))
65#define MIN_WORD_LEN (4)
66
67/* SPI_TX/SPI_RX_TRANS_LEN fields */
68#define TRANS_LEN_MSK (GENMASK(23, 0))
69
70/* M_CMD OP codes for SPI */
71#define SPI_TX_ONLY (1)
72#define SPI_RX_ONLY (2)
73#define SPI_FULL_DUPLEX (3)
74#define SPI_TX_RX (7)
75#define SPI_CS_ASSERT (8)
76#define SPI_CS_DEASSERT (9)
77#define SPI_SCK_ONLY (10)
78/* M_CMD params for SPI */
Girish Mahadevan061b9e32017-05-18 10:25:43 -060079#define SPI_PRE_CMD_DELAY BIT(0)
80#define TIMESTAMP_BEFORE BIT(1)
81#define FRAGMENTATION BIT(2)
82#define TIMESTAMP_AFTER BIT(3)
83#define POST_CMD_DELAY BIT(4)
Girish Mahadevan2ef85af2017-02-14 14:42:22 -070084
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -060085#define SPI_CORE2X_VOTE (10000)
Girish Mahadevan9a7c9442017-08-15 12:10:09 -060086/* GSI CONFIG0 TRE Params */
87/* Flags bit fields */
88#define GSI_LOOPBACK_EN (BIT(0))
89#define GSI_CS_TOGGLE (BIT(3))
90#define GSI_CPHA (BIT(4))
91#define GSI_CPOL (BIT(5))
92
93#define MAX_TX_SG (3)
94#define NUM_SPI_XFER (8)
95
96struct gsi_desc_cb {
97 struct spi_master *spi;
98 struct spi_transfer *xfer;
99};
100
101struct spi_geni_gsi {
102 struct msm_gpi_tre config0_tre;
103 struct msm_gpi_tre go_tre;
104 struct msm_gpi_tre tx_dma_tre;
105 struct msm_gpi_tre rx_dma_tre;
106 struct scatterlist tx_sg[MAX_TX_SG];
107 struct scatterlist rx_sg;
108 dma_cookie_t tx_cookie;
109 dma_cookie_t rx_cookie;
110 struct msm_gpi_dma_async_tx_cb_param tx_cb_param;
111 struct msm_gpi_dma_async_tx_cb_param rx_cb_param;
112 struct dma_async_tx_descriptor *tx_desc;
113 struct dma_async_tx_descriptor *rx_desc;
114 struct gsi_desc_cb desc_cb;
115};
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -0600116
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700117struct spi_geni_master {
118 struct se_geni_rsc spi_rsc;
119 resource_size_t phys_addr;
120 resource_size_t size;
121 void __iomem *base;
122 int irq;
123 struct device *dev;
124 int rx_fifo_depth;
125 int tx_fifo_depth;
126 int tx_fifo_width;
127 int tx_wm;
128 bool setup;
129 u32 cur_speed_hz;
130 int cur_word_len;
131 unsigned int tx_rem_bytes;
132 unsigned int rx_rem_bytes;
133 struct spi_transfer *cur_xfer;
134 struct completion xfer_done;
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -0600135 struct device *wrapper_dev;
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600136 int oversampling;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600137 struct spi_geni_gsi *gsi;
138 struct dma_chan *tx;
139 struct dma_chan *rx;
140 struct msm_gpi_ctrl tx_event;
141 struct msm_gpi_ctrl rx_event;
142 struct completion tx_cb;
143 struct completion rx_cb;
144 bool qn_err;
145 int cur_xfer_mode;
146 int num_tx_eot;
147 int num_rx_eot;
148 int num_xfers;
149 void *ipc;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700150};
151
152static struct spi_master *get_spi_master(struct device *dev)
153{
154 struct platform_device *pdev = to_platform_device(dev);
155 struct spi_master *spi = platform_get_drvdata(pdev);
156
157 return spi;
158}
159
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600160static int get_spi_clk_cfg(u32 speed_hz, struct spi_geni_master *mas,
161 int *clk_idx, int *clk_div)
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700162{
163 unsigned long sclk_freq;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700164 struct se_geni_rsc *rsc = &mas->spi_rsc;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600165 int ret = 0;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700166
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600167 ret = geni_se_clk_freq_match(&mas->spi_rsc,
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600168 (speed_hz * mas->oversampling), clk_idx,
169 &sclk_freq, true);
Girish Mahadevan13f3f002017-06-05 14:27:57 -0600170 if (ret) {
171 dev_err(mas->dev, "%s: Failed(%d) to find src clk for 0x%x\n",
172 __func__, ret, speed_hz);
173 return ret;
174 }
Girish Mahadevan6727acc2017-04-05 12:40:19 -0600175
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600176 *clk_div = ((sclk_freq / mas->oversampling) / speed_hz);
177 if (!(*clk_div)) {
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600178 dev_err(mas->dev, "%s:Err:sclk:%lu oversampling:%d speed:%u\n",
179 __func__, sclk_freq, mas->oversampling, speed_hz);
Girish Mahadevan6727acc2017-04-05 12:40:19 -0600180 return -EINVAL;
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600181 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700182
Girish Mahadevan13f3f002017-06-05 14:27:57 -0600183 dev_dbg(mas->dev, "%s: req %u sclk %lu, idx %d, div %d\n", __func__,
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600184 speed_hz, sclk_freq, *clk_idx, *clk_div);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700185 ret = clk_set_rate(rsc->se_clk, sclk_freq);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600186 if (ret)
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600187 dev_err(mas->dev, "%s: clk_set_rate failed %d\n",
188 __func__, ret);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600189 return ret;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700190}
191
192static void spi_setup_word_len(struct spi_geni_master *mas, u32 mode,
193 int bits_per_word)
194{
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600195 int pack_words = 1;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700196 bool msb_first = (mode & SPI_LSB_FIRST) ? false : true;
197 u32 word_len = geni_read_reg(mas->base, SE_SPI_WORD_LEN);
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600198 unsigned long cfg0, cfg1;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700199
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600200 /*
201 * If bits_per_word isn't a byte aligned value, set the packing to be
202 * 1 SPI word per FIFO word.
203 */
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600204 if (!(mas->tx_fifo_width % bits_per_word))
205 pack_words = mas->tx_fifo_width / bits_per_word;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700206 word_len &= ~WORD_LEN_MSK;
207 word_len |= ((bits_per_word - MIN_WORD_LEN) & WORD_LEN_MSK);
208 se_config_packing(mas->base, bits_per_word, pack_words, msb_first);
209 geni_write_reg(word_len, mas->base, SE_SPI_WORD_LEN);
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600210 se_get_packing_config(bits_per_word, pack_words, msb_first,
211 &cfg0, &cfg1);
212 GENI_SE_DBG(mas->ipc, false, mas->dev,
213 "%s: cfg0 %lu cfg1 %lu bpw %d pack_words %d\n", __func__,
214 cfg0, cfg1, bits_per_word, pack_words);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700215}
216
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600217static int setup_fifo_params(struct spi_device *spi_slv,
218 struct spi_master *spi)
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700219{
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600220 struct spi_geni_master *mas = spi_master_get_devdata(spi);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700221 u16 mode = spi_slv->mode;
222 u32 loopback_cfg = geni_read_reg(mas->base, SE_SPI_LOOPBACK);
223 u32 cpol = geni_read_reg(mas->base, SE_SPI_CPOL);
224 u32 cpha = geni_read_reg(mas->base, SE_SPI_CPHA);
Girish Mahadevan061b9e32017-05-18 10:25:43 -0600225 u32 demux_sel = 0;
226 u32 demux_output_inv = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600227 u32 clk_sel = geni_read_reg(mas->base, SE_GENI_CLK_SEL);
228 u32 m_clk_cfg = geni_read_reg(mas->base, GENI_SER_M_CLK_CFG);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700229 int ret = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600230 int idx;
231 int div;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700232
233 loopback_cfg &= ~LOOPBACK_MSK;
234 cpol &= ~CPOL;
235 cpha &= ~CPHA;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700236
237 if (mode & SPI_LOOP)
238 loopback_cfg |= LOOPBACK_ENABLE;
239
240 if (mode & SPI_CPOL)
241 cpol |= CPOL;
242
243 if (mode & SPI_CPHA)
244 cpha |= CPHA;
245
246 if (spi_slv->mode & SPI_CS_HIGH)
247 demux_output_inv |= BIT(spi_slv->chip_select);
248
Girish Mahadevan061b9e32017-05-18 10:25:43 -0600249 demux_sel = spi_slv->chip_select;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700250 mas->cur_speed_hz = spi_slv->max_speed_hz;
251 mas->cur_word_len = spi_slv->bits_per_word;
252
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600253 ret = get_spi_clk_cfg(mas->cur_speed_hz, mas, &idx, &div);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700254 if (ret) {
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600255 dev_err(mas->dev, "Err setting clks ret(%d) for %d\n",
Girish Mahadevan6727acc2017-04-05 12:40:19 -0600256 ret, mas->cur_speed_hz);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600257 goto setup_fifo_params_exit;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700258 }
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600259
260 clk_sel |= (idx & CLK_SEL_MSK);
261 m_clk_cfg |= ((div << CLK_DIV_SHFT) | SER_CLK_EN);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700262 spi_setup_word_len(mas, spi_slv->mode, spi_slv->bits_per_word);
263 geni_write_reg(loopback_cfg, mas->base, SE_SPI_LOOPBACK);
264 geni_write_reg(demux_sel, mas->base, SE_SPI_DEMUX_SEL);
265 geni_write_reg(cpha, mas->base, SE_SPI_CPHA);
266 geni_write_reg(cpol, mas->base, SE_SPI_CPOL);
267 geni_write_reg(demux_output_inv, mas->base, SE_SPI_DEMUX_OUTPUT_INV);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600268 geni_write_reg(clk_sel, mas->base, SE_GENI_CLK_SEL);
269 geni_write_reg(m_clk_cfg, mas->base, GENI_SER_M_CLK_CFG);
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600270 GENI_SE_DBG(mas->ipc, false, mas->dev,
271 "%s:Loopback%d demux_sel0x%x demux_op_inv 0x%x clk_cfg 0x%x\n",
272 __func__, loopback_cfg, demux_sel, demux_output_inv, m_clk_cfg);
273 GENI_SE_DBG(mas->ipc, false, mas->dev,
274 "%s:clk_sel 0x%x cpol %d cpha %d\n", __func__,
275 clk_sel, cpol, cpha);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700276 /* Ensure message level attributes are written before returning */
277 mb();
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600278setup_fifo_params_exit:
279 return ret;
280}
281
282
283static int select_xfer_mode(struct spi_master *spi,
284 struct spi_message *spi_msg)
285{
286 struct spi_geni_master *mas = spi_master_get_devdata(spi);
287 int mode = FIFO_MODE;
288 int fifo_disable = (geni_read_reg(mas->base, GENI_IF_FIFO_DISABLE_RO) &
289 FIFO_IF_DISABLE);
290 bool dma_chan_valid =
291 !(IS_ERR_OR_NULL(mas->tx) || IS_ERR_OR_NULL(mas->rx));
292
293 /*
294 * If FIFO Interface is disabled and there are no DMA channels then we
295 * can't do this transfer.
296 * If FIFO interface is disabled, we can do GSI only,
297 * else pick FIFO mode.
298 */
299 if (fifo_disable && !dma_chan_valid)
300 mode = -EINVAL;
301 else if (fifo_disable)
302 mode = GSI_DMA;
303 else
304 mode = FIFO_MODE;
305 return mode;
306}
307
308static struct msm_gpi_tre *setup_config0_tre(struct spi_transfer *xfer,
309 struct spi_geni_master *mas, u16 mode)
310{
311 struct msm_gpi_tre *c0_tre = &mas->gsi[mas->num_xfers].config0_tre;
312 u8 flags = 0;
313 u8 word_len = 0;
314 u8 pack = 0;
315 int div = 0;
316 int idx = 0;
317 int ret = 0;
318
319 if (IS_ERR_OR_NULL(c0_tre))
320 return c0_tre;
321
322 if (mode & SPI_LOOP)
323 flags |= GSI_LOOPBACK_EN;
324
325 if (mode & SPI_CPOL)
326 flags |= GSI_CPOL;
327
328 if (mode & SPI_CPHA)
329 flags |= GSI_CPHA;
330
331 if (xfer->cs_change)
332 flags |= GSI_CS_TOGGLE;
333
334 word_len = xfer->bits_per_word - MIN_WORD_LEN;
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600335 pack |= (GSI_TX_PACK_EN | GSI_RX_PACK_EN);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600336 ret = get_spi_clk_cfg(mas->cur_speed_hz, mas, &idx, &div);
337 if (ret) {
338 dev_err(mas->dev, "%s:Err setting clks:%d\n", __func__, ret);
339 return ERR_PTR(ret);
340 }
341 c0_tre->dword[0] = MSM_GPI_SPI_CONFIG0_TRE_DWORD0(pack, flags,
342 word_len);
343 c0_tre->dword[1] = MSM_GPI_SPI_CONFIG0_TRE_DWORD1(0, 0, 0);
344 c0_tre->dword[2] = MSM_GPI_SPI_CONFIG0_TRE_DWORD2(idx, div);
345 c0_tre->dword[3] = MSM_GPI_SPI_CONFIG0_TRE_DWORD3(0, 0, 0, 1);
346 GENI_SE_DBG(mas->ipc, false, mas->dev,
347 "%s: flags 0x%x word %d pack %d idx %d div %d\n",
348 __func__, flags, word_len, pack, idx, div);
349 return c0_tre;
350}
351
352static struct msm_gpi_tre *setup_go_tre(int cmd, int cs, int rx_len, int flags,
353 struct spi_geni_master *mas)
354{
355 struct msm_gpi_tre *go_tre = &mas->gsi[mas->num_xfers].go_tre;
356 int chain;
357 int eot;
358 int eob;
359
360 if (IS_ERR_OR_NULL(go_tre))
361 return go_tre;
362
363 go_tre->dword[0] = MSM_GPI_SPI_GO_TRE_DWORD0(flags, cs, cmd);
364 go_tre->dword[1] = MSM_GPI_SPI_GO_TRE_DWORD1;
365 go_tre->dword[2] = MSM_GPI_SPI_GO_TRE_DWORD2(rx_len);
366 if (cmd == SPI_RX_ONLY) {
367 eot = 0;
368 chain = 0;
369 eob = 1;
370 } else {
371 eot = 0;
372 chain = 1;
373 eob = 0;
374 }
375 go_tre->dword[3] = MSM_GPI_SPI_GO_TRE_DWORD3(0, eot, eob, chain);
376 GENI_SE_DBG(mas->ipc, false, mas->dev,
377 "%s: rx len %d flags 0x%x cs %d cmd %d eot %d eob %d chain %d\n",
378 __func__, rx_len, flags, cs, cmd, eot, eob, chain);
379 return go_tre;
380}
381
382static struct msm_gpi_tre *setup_dma_tre(struct msm_gpi_tre *tre,
383 dma_addr_t buf, u32 len,
384 struct spi_geni_master *mas,
385 bool is_tx)
386{
387 if (IS_ERR_OR_NULL(tre))
388 return tre;
389
390 tre->dword[0] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD0(buf);
391 tre->dword[1] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD1(buf);
392 tre->dword[2] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD2(len);
393 tre->dword[3] = MSM_GPI_DMA_W_BUFFER_TRE_DWORD3(0, is_tx, 0, 0);
394 return tre;
395}
396
397static void spi_gsi_ch_cb(struct dma_chan *ch, struct msm_gpi_cb const *cb,
398 void *ptr)
399{
400 struct spi_master *spi = ptr;
401 struct spi_geni_master *mas = spi_master_get_devdata(spi);
402
403 switch (cb->cb_event) {
404 case MSM_GPI_QUP_NOTIFY:
405 case MSM_GPI_QUP_MAX_EVENT:
406 GENI_SE_DBG(mas->ipc, false, mas->dev,
407 "%s:cb_ev%d status%llu ts%llu count%llu\n",
408 __func__, cb->cb_event, cb->status,
409 cb->timestamp, cb->count);
410 break;
411 case MSM_GPI_QUP_ERROR:
412 case MSM_GPI_QUP_CH_ERROR:
413 case MSM_GPI_QUP_FW_ERROR:
414 case MSM_GPI_QUP_PENDING_EVENT:
415 case MSM_GPI_QUP_EOT_DESC_MISMATCH:
416 case MSM_GPI_QUP_SW_ERROR:
417 GENI_SE_ERR(mas->ipc, true, mas->dev,
418 "%s: cb_ev %d status %llu ts %llu count %llu\n",
419 __func__, cb->cb_event, cb->status,
420 cb->timestamp, cb->count);
421 GENI_SE_ERR(mas->ipc, true, mas->dev,
422 "err.routine %u, err.type %u, err.code %u\n",
423 cb->error_log.routine,
424 cb->error_log.type,
425 cb->error_log.error_code);
426 mas->qn_err = true;
427 complete_all(&mas->tx_cb);
428 complete_all(&mas->rx_cb);
429
430 break;
431 };
432}
433
434static void spi_gsi_rx_callback(void *cb)
435{
436 struct msm_gpi_dma_async_tx_cb_param *cb_param =
437 (struct msm_gpi_dma_async_tx_cb_param *)cb;
438 struct gsi_desc_cb *desc_cb = (struct gsi_desc_cb *)cb_param->userdata;
439 struct spi_master *spi = desc_cb->spi;
440 struct spi_transfer *xfer = desc_cb->xfer;
441 struct spi_geni_master *mas = spi_master_get_devdata(spi);
442
443 if (xfer->rx_buf) {
444 if (cb_param->status == MSM_GPI_TCE_UNEXP_ERR) {
445 GENI_SE_ERR(mas->ipc, true, mas->dev,
446 "%s: Unexpected GSI CB error\n", __func__);
447 return;
448 }
449 if (cb_param->length == xfer->len) {
450 GENI_SE_DBG(mas->ipc, false, mas->dev,
451 "%s\n", __func__);
452 complete(&mas->rx_cb);
453 } else {
454 GENI_SE_ERR(mas->ipc, true, mas->dev,
455 "%s: Length mismatch. Expected %d Callback %d\n",
456 __func__, xfer->len, cb_param->length);
457 }
458 }
459}
460
461static void spi_gsi_tx_callback(void *cb)
462{
463 struct msm_gpi_dma_async_tx_cb_param *cb_param = cb;
464 struct gsi_desc_cb *desc_cb = (struct gsi_desc_cb *)cb_param->userdata;
465 struct spi_master *spi = desc_cb->spi;
466 struct spi_transfer *xfer = desc_cb->xfer;
467 struct spi_geni_master *mas = spi_master_get_devdata(spi);
468
469 if (xfer->tx_buf) {
470 if (cb_param->status == MSM_GPI_TCE_UNEXP_ERR) {
471 GENI_SE_ERR(mas->ipc, true, mas->dev,
472 "%s: Unexpected GSI CB error\n", __func__);
473 return;
474 }
475 if (cb_param->length == xfer->len) {
476 GENI_SE_DBG(mas->ipc, false, mas->dev,
477 "%s\n", __func__);
478 complete(&mas->tx_cb);
479 } else {
480 GENI_SE_ERR(mas->ipc, true, mas->dev,
481 "%s: Length mismatch. Expected %d Callback %d\n",
482 __func__, xfer->len, cb_param->length);
483 }
484 }
485}
486
487static int setup_gsi_xfer(struct spi_transfer *xfer,
488 struct spi_geni_master *mas,
489 struct spi_device *spi_slv,
490 struct spi_master *spi)
491{
492 int ret = 0;
493 struct msm_gpi_tre *c0_tre = NULL;
494 struct msm_gpi_tre *go_tre = NULL;
495 struct msm_gpi_tre *tx_tre = NULL;
496 struct msm_gpi_tre *rx_tre = NULL;
497 struct scatterlist *xfer_tx_sg = mas->gsi[mas->num_xfers].tx_sg;
498 struct scatterlist *xfer_rx_sg = &mas->gsi[mas->num_xfers].rx_sg;
499 int rx_nent = 0;
500 int tx_nent = 0;
501 u8 cmd = 0;
502 u8 cs = 0;
503 u32 rx_len = 0;
504 int go_flags = 0;
505 unsigned long flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK;
506
507 if ((xfer->bits_per_word != mas->cur_word_len) ||
508 (xfer->speed_hz != mas->cur_speed_hz)) {
509 mas->cur_word_len = xfer->bits_per_word;
510 mas->cur_speed_hz = xfer->speed_hz;
511 tx_nent++;
512 c0_tre = setup_config0_tre(xfer, mas, spi_slv->mode);
513 if (IS_ERR_OR_NULL(c0_tre)) {
514 dev_err(mas->dev, "%s:Err setting c0tre:%d\n",
515 __func__, ret);
516 return PTR_ERR(c0_tre);
517 }
518 }
519
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600520 if (!(mas->cur_word_len % MIN_WORD_LEN)) {
521 rx_len = ((xfer->len << 3) / mas->cur_word_len);
522 } else {
523 int bytes_per_word = (mas->cur_word_len / BITS_PER_BYTE) + 1;
524
525 rx_len = (xfer->len / bytes_per_word);
526 }
527
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600528 if (xfer->tx_buf && xfer->rx_buf) {
529 cmd = SPI_FULL_DUPLEX;
530 tx_nent += 2;
531 rx_nent++;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600532 } else if (xfer->tx_buf) {
533 cmd = SPI_TX_ONLY;
534 tx_nent += 2;
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600535 rx_len = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600536 } else if (xfer->rx_buf) {
537 cmd = SPI_RX_ONLY;
538 tx_nent++;
539 rx_nent++;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600540 }
541
542 cs |= spi_slv->chip_select;
543 if (!list_is_last(&xfer->transfer_list, &spi->cur_msg->transfers))
544 go_flags |= FRAGMENTATION;
545 go_tre = setup_go_tre(cmd, cs, rx_len, go_flags, mas);
546
547 sg_init_table(xfer_tx_sg, tx_nent);
548 if (rx_nent)
549 sg_init_table(xfer_rx_sg, rx_nent);
550
551 if (c0_tre)
552 sg_set_buf(xfer_tx_sg++, c0_tre, sizeof(*c0_tre));
553
554 sg_set_buf(xfer_tx_sg++, go_tre, sizeof(*go_tre));
555 mas->gsi[mas->num_xfers].desc_cb.spi = spi;
556 mas->gsi[mas->num_xfers].desc_cb.xfer = xfer;
557 if (cmd & SPI_RX_ONLY) {
558 rx_tre = &mas->gsi[mas->num_xfers].rx_dma_tre;
559 rx_tre = setup_dma_tre(rx_tre, xfer->rx_dma, xfer->len, mas, 0);
560 if (IS_ERR_OR_NULL(rx_tre)) {
561 dev_err(mas->dev, "Err setting up rx tre\n");
562 return PTR_ERR(rx_tre);
563 }
564 sg_set_buf(xfer_rx_sg, rx_tre, sizeof(*rx_tre));
565 mas->gsi[mas->num_xfers].rx_desc =
566 dmaengine_prep_slave_sg(mas->rx,
567 &mas->gsi[mas->num_xfers].rx_sg, rx_nent,
568 DMA_DEV_TO_MEM, flags);
569 if (IS_ERR_OR_NULL(mas->gsi[mas->num_xfers].rx_desc)) {
570 dev_err(mas->dev, "Err setting up rx desc\n");
571 return -EIO;
572 }
573 mas->gsi[mas->num_xfers].rx_desc->callback =
574 spi_gsi_rx_callback;
575 mas->gsi[mas->num_xfers].rx_desc->callback_param =
576 &mas->gsi[mas->num_xfers].rx_cb_param;
577 mas->gsi[mas->num_xfers].rx_cb_param.userdata =
578 &mas->gsi[mas->num_xfers].desc_cb;
579 mas->num_rx_eot++;
580 }
581
582 if (cmd & SPI_TX_ONLY) {
583 tx_tre = &mas->gsi[mas->num_xfers].tx_dma_tre;
584 tx_tre = setup_dma_tre(tx_tre, xfer->tx_dma, xfer->len, mas, 1);
585 if (IS_ERR_OR_NULL(tx_tre)) {
586 dev_err(mas->dev, "Err setting up tx tre\n");
587 return PTR_ERR(tx_tre);
588 }
589 sg_set_buf(xfer_tx_sg++, tx_tre, sizeof(*tx_tre));
590 mas->num_tx_eot++;
591 }
592 mas->gsi[mas->num_xfers].tx_desc = dmaengine_prep_slave_sg(mas->tx,
593 mas->gsi[mas->num_xfers].tx_sg, tx_nent,
594 DMA_MEM_TO_DEV, flags);
595 if (IS_ERR_OR_NULL(mas->gsi[mas->num_xfers].tx_desc)) {
596 dev_err(mas->dev, "Err setting up tx desc\n");
597 return -EIO;
598 }
599 mas->gsi[mas->num_xfers].tx_desc->callback = spi_gsi_tx_callback;
600 mas->gsi[mas->num_xfers].tx_desc->callback_param =
601 &mas->gsi[mas->num_xfers].tx_cb_param;
602 mas->gsi[mas->num_xfers].tx_cb_param.userdata =
603 &mas->gsi[mas->num_xfers].desc_cb;
604 mas->gsi[mas->num_xfers].tx_cookie =
605 dmaengine_submit(mas->gsi[mas->num_xfers].tx_desc);
606 if (mas->num_rx_eot)
607 mas->gsi[mas->num_xfers].rx_cookie =
608 dmaengine_submit(mas->gsi[mas->num_xfers].rx_desc);
609 dma_async_issue_pending(mas->tx);
610 if (mas->num_rx_eot)
611 dma_async_issue_pending(mas->rx);
612 mas->num_xfers++;
613 return ret;
614}
615
616static int spi_geni_map_buf(struct spi_geni_master *mas,
617 struct spi_message *msg)
618{
619 struct spi_transfer *xfer;
620 struct device *gsi_dev = mas->dev;
621
622 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
623 if (xfer->rx_buf) {
624 xfer->rx_dma = dma_map_single(gsi_dev, xfer->rx_buf,
625 xfer->len, DMA_FROM_DEVICE);
626 if (dma_mapping_error(mas->dev, xfer->rx_dma)) {
627 dev_err(mas->dev, "Err mapping buf\n");
628 return -ENOMEM;
629 }
630 }
631
632 if (xfer->tx_buf) {
633 xfer->tx_dma = dma_map_single(gsi_dev,
634 (void *)xfer->tx_buf, xfer->len, DMA_TO_DEVICE);
635 if (dma_mapping_error(gsi_dev, xfer->tx_dma)) {
636 dev_err(mas->dev, "Err mapping buf\n");
637 dma_unmap_single(gsi_dev, xfer->rx_dma,
638 xfer->len, DMA_FROM_DEVICE);
639 return -ENOMEM;
640 }
641 }
642 };
643
644 return 0;
645}
646
647static void spi_geni_unmap_buf(struct spi_geni_master *mas,
648 struct spi_message *msg)
649{
650 struct spi_transfer *xfer;
651 struct device *gsi_dev = mas->dev;
652
653 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
654 if (xfer->rx_buf)
655 dma_unmap_single(gsi_dev, xfer->rx_dma,
656 xfer->len, DMA_FROM_DEVICE);
657 if (xfer->tx_buf)
658 dma_unmap_single(gsi_dev, xfer->tx_dma,
659 xfer->len, DMA_TO_DEVICE);
660 };
661}
662
663static int spi_geni_prepare_message(struct spi_master *spi,
664 struct spi_message *spi_msg)
665{
666 int ret = 0;
667 struct spi_geni_master *mas = spi_master_get_devdata(spi);
668
669 mas->cur_xfer_mode = select_xfer_mode(spi, spi_msg);
670
671 if (mas->cur_xfer_mode == FIFO_MODE) {
672 geni_se_select_mode(mas->base, FIFO_MODE);
673 reinit_completion(&mas->xfer_done);
674 setup_fifo_params(spi_msg->spi, spi);
675 } else if (mas->cur_xfer_mode == GSI_DMA) {
676 mas->num_tx_eot = 0;
677 mas->num_rx_eot = 0;
678 mas->num_xfers = 0;
679 reinit_completion(&mas->tx_cb);
680 reinit_completion(&mas->rx_cb);
681 memset(mas->gsi, 0,
682 (sizeof(struct spi_geni_gsi) * NUM_SPI_XFER));
683 geni_se_select_mode(mas->base, GSI_DMA);
684 dmaengine_resume(mas->tx);
685 ret = spi_geni_map_buf(mas, spi_msg);
686 } else {
687 dev_err(mas->dev, "%s: Couldn't select mode %d", __func__,
688 mas->cur_xfer_mode);
689 ret = -EINVAL;
690 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700691 return ret;
692}
693
694static int spi_geni_unprepare_message(struct spi_master *spi_mas,
695 struct spi_message *spi_msg)
696{
697 struct spi_geni_master *mas = spi_master_get_devdata(spi_mas);
698
699 mas->cur_speed_hz = 0;
700 mas->cur_word_len = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600701 if (mas->cur_xfer_mode == GSI_DMA) {
702 dmaengine_pause(mas->tx);
703 spi_geni_unmap_buf(mas, spi_msg);
704 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700705 return 0;
706}
707
708static int spi_geni_prepare_transfer_hardware(struct spi_master *spi)
709{
710 struct spi_geni_master *mas = spi_master_get_devdata(spi);
711 int ret = 0;
712
713 ret = pm_runtime_get_sync(mas->dev);
714 if (ret < 0) {
715 dev_err(mas->dev, "Error enabling SE resources\n");
716 pm_runtime_put_noidle(mas->dev);
717 goto exit_prepare_transfer_hardware;
718 } else {
719 ret = 0;
720 }
721
722 if (unlikely(!mas->setup)) {
723 int proto = get_se_proto(mas->base);
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600724 unsigned int major;
725 unsigned int minor;
726 unsigned int step;
727 int hw_ver;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700728
729 if (unlikely(proto != SPI)) {
730 dev_err(mas->dev, "Invalid proto %d\n", proto);
731 return -ENXIO;
732 }
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -0600733 geni_se_init(mas->base, 0x0, (mas->tx_fifo_depth - 2));
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700734 mas->tx_fifo_depth = get_tx_fifo_depth(mas->base);
735 mas->rx_fifo_depth = get_rx_fifo_depth(mas->base);
736 mas->tx_fifo_width = get_tx_fifo_width(mas->base);
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600737 mas->oversampling = 1;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700738 /* Transmit an entire FIFO worth of data per IRQ */
739 mas->tx_wm = 1;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600740
741 mas->tx = dma_request_slave_channel(mas->dev, "tx");
742 if (IS_ERR_OR_NULL(mas->tx)) {
743 dev_info(mas->dev, "Failed to get tx DMA ch %ld",
744 PTR_ERR(mas->tx));
745 } else {
746 mas->rx = dma_request_slave_channel(mas->dev, "rx");
747 if (IS_ERR_OR_NULL(mas->rx)) {
748 dev_info(mas->dev, "Failed to get rx DMA ch %ld",
749 PTR_ERR(mas->rx));
750 dma_release_channel(mas->tx);
751 }
752 mas->gsi = devm_kzalloc(mas->dev,
753 (sizeof(struct spi_geni_gsi) * NUM_SPI_XFER),
754 GFP_KERNEL);
755 if (IS_ERR_OR_NULL(mas->gsi)) {
756 dev_err(mas->dev, "Failed to get GSI mem\n");
757 dma_release_channel(mas->tx);
758 dma_release_channel(mas->rx);
759 mas->tx = NULL;
760 mas->rx = NULL;
761 goto setup_ipc;
762 }
763 mas->tx_event.init.callback = spi_gsi_ch_cb;
764 mas->tx_event.init.cb_param = spi;
765 mas->tx_event.cmd = MSM_GPI_INIT;
766 mas->tx->private = &mas->tx_event;
767 mas->rx_event.init.callback = spi_gsi_ch_cb;
768 mas->rx_event.init.cb_param = spi;
769 mas->rx_event.cmd = MSM_GPI_INIT;
770 mas->rx->private = &mas->rx_event;
771 if (dmaengine_slave_config(mas->tx, NULL)) {
772 dev_err(mas->dev, "Failed to Config Tx\n");
773 dma_release_channel(mas->tx);
774 dma_release_channel(mas->rx);
775 mas->tx = NULL;
776 mas->rx = NULL;
777 goto setup_ipc;
778 }
779 if (dmaengine_slave_config(mas->rx, NULL)) {
780 dev_err(mas->dev, "Failed to Config Rx\n");
781 dma_release_channel(mas->tx);
782 dma_release_channel(mas->rx);
783 mas->tx = NULL;
784 mas->rx = NULL;
785 goto setup_ipc;
786 }
787
788 }
789setup_ipc:
790 mas->ipc = ipc_log_context_create(4, dev_name(mas->dev), 0);
791 dev_info(mas->dev, "tx_fifo %d rx_fifo %d tx_width %d\n",
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700792 mas->tx_fifo_depth, mas->rx_fifo_depth,
793 mas->tx_fifo_width);
794 mas->setup = true;
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600795 hw_ver = geni_se_qupv3_hw_version(mas->wrapper_dev, &major,
796 &minor, &step);
797 if (hw_ver)
798 dev_err(mas->dev, "%s:Err getting HW version %d\n",
799 __func__, hw_ver);
800 else {
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600801 if ((major == 1) && (minor == 0))
802 mas->oversampling = 2;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600803 GENI_SE_DBG(mas->ipc, false, mas->dev,
804 "%s:Major:%d Minor:%d step:%dos%d\n",
805 __func__, major, minor, step, mas->oversampling);
Girish Mahadevan96cf38f2017-07-31 11:01:55 -0600806 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700807 }
808exit_prepare_transfer_hardware:
809 return ret;
810}
811
812static int spi_geni_unprepare_transfer_hardware(struct spi_master *spi)
813{
814 struct spi_geni_master *mas = spi_master_get_devdata(spi);
815
816 pm_runtime_put_sync(mas->dev);
817 return 0;
818}
819
820static void setup_fifo_xfer(struct spi_transfer *xfer,
821 struct spi_geni_master *mas, u16 mode,
822 struct spi_master *spi)
823{
824 u32 m_cmd = 0;
825 u32 m_param = 0;
826 u32 spi_tx_cfg = geni_read_reg(mas->base, SE_SPI_TRANS_CFG);
827 u32 trans_len = 0;
828
829 if (xfer->bits_per_word != mas->cur_word_len) {
830 spi_setup_word_len(mas, mode, xfer->bits_per_word);
831 mas->cur_word_len = xfer->bits_per_word;
832 }
833
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600834 /* Speed and bits per word can be overridden per transfer */
835 if (xfer->speed_hz != mas->cur_speed_hz) {
836 int ret = 0;
837 u32 clk_sel = geni_read_reg(mas->base, SE_GENI_CLK_SEL);
838 u32 m_clk_cfg = geni_read_reg(mas->base, GENI_SER_M_CLK_CFG);
839 int idx = 0;
840 int div = 0;
841
842 ret = get_spi_clk_cfg(xfer->speed_hz, mas, &idx, &div);
843 if (ret) {
844 dev_err(mas->dev, "%s:Err setting clks:%d\n",
845 __func__, ret);
846 return;
847 }
848 mas->cur_speed_hz = xfer->speed_hz;
849 clk_sel |= (idx & CLK_SEL_MSK);
850 m_clk_cfg |= ((div << CLK_DIV_SHFT) | SER_CLK_EN);
851 geni_write_reg(clk_sel, mas->base, SE_GENI_CLK_SEL);
852 geni_write_reg(m_clk_cfg, mas->base, GENI_SER_M_CLK_CFG);
853 }
854
855 mas->tx_rem_bytes = 0;
856 mas->rx_rem_bytes = 0;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700857 if (xfer->tx_buf && xfer->rx_buf)
858 m_cmd = SPI_FULL_DUPLEX;
859 else if (xfer->tx_buf)
860 m_cmd = SPI_TX_ONLY;
861 else if (xfer->rx_buf)
862 m_cmd = SPI_RX_ONLY;
863
864 spi_tx_cfg &= ~CS_TOGGLE;
865 if (xfer->cs_change)
866 spi_tx_cfg |= CS_TOGGLE;
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600867 if (!(mas->cur_word_len % MIN_WORD_LEN)) {
868 trans_len =
869 ((xfer->len << 3) / mas->cur_word_len) & TRANS_LEN_MSK;
870 } else {
871 int bytes_per_word = (mas->cur_word_len / BITS_PER_BYTE) + 1;
872
873 trans_len = (xfer->len / bytes_per_word) & TRANS_LEN_MSK;
874 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700875 if (!list_is_last(&xfer->transfer_list, &spi->cur_msg->transfers))
876 m_param |= FRAGMENTATION;
877
878 mas->cur_xfer = xfer;
879 if (m_cmd & SPI_TX_ONLY) {
880 mas->tx_rem_bytes = xfer->len;
881 geni_write_reg(trans_len, mas->base, SE_SPI_TX_TRANS_LEN);
882 }
883
884 if (m_cmd & SPI_RX_ONLY) {
885 geni_write_reg(trans_len, mas->base, SE_SPI_RX_TRANS_LEN);
886 mas->rx_rem_bytes = xfer->len;
887 }
888 geni_write_reg(spi_tx_cfg, mas->base, SE_SPI_TRANS_CFG);
889 geni_setup_m_cmd(mas->base, m_cmd, m_param);
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600890 GENI_SE_DBG(mas->ipc, false, mas->dev,
891 "%s: trans_len %d xferlen%d tx_cfg 0x%x cmd 0x%x\n",
892 __func__, trans_len, xfer->len, spi_tx_cfg, m_cmd);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600893 if (m_cmd & SPI_TX_ONLY)
894 geni_write_reg(mas->tx_wm, mas->base, SE_GENI_TX_WATERMARK_REG);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700895 /* Ensure all writes are done before the WM interrupt */
896 mb();
897}
898
899static void handle_fifo_timeout(struct spi_geni_master *mas)
900{
901 unsigned long timeout;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700902
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700903 reinit_completion(&mas->xfer_done);
904 geni_cancel_m_cmd(mas->base);
Girish Mahadevan1fc27702017-08-31 12:55:23 -0600905 geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700906 /* Ensure cmd cancel is written */
907 mb();
908 timeout = wait_for_completion_timeout(&mas->xfer_done, HZ);
909 if (!timeout) {
910 reinit_completion(&mas->xfer_done);
911 geni_abort_m_cmd(mas->base);
912 /* Ensure cmd abort is written */
913 mb();
914 timeout = wait_for_completion_timeout(&mas->xfer_done,
915 HZ);
916 if (!timeout)
917 dev_err(mas->dev,
918 "Failed to cancel/abort m_cmd\n");
919 }
920}
921
922static int spi_geni_transfer_one(struct spi_master *spi,
923 struct spi_device *slv,
924 struct spi_transfer *xfer)
925{
926 int ret = 0;
927 struct spi_geni_master *mas = spi_master_get_devdata(spi);
928 unsigned long timeout;
929
930 if ((xfer->tx_buf == NULL) && (xfer->rx_buf == NULL)) {
931 dev_err(mas->dev, "Invalid xfer both tx rx are NULL\n");
932 return -EINVAL;
933 }
934
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600935 if (mas->cur_xfer_mode == FIFO_MODE) {
936 setup_fifo_xfer(xfer, mas, slv->mode, spi);
937 timeout = wait_for_completion_timeout(&mas->xfer_done,
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700938 msecs_to_jiffies(SPI_XFER_TIMEOUT_MS));
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600939 if (!timeout) {
940 GENI_SE_ERR(mas->ipc, true, mas->dev,
941 "Xfer[len %d tx %pK rx %pK n %d] timed out.\n",
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700942 xfer->len, xfer->tx_buf,
943 xfer->rx_buf,
944 xfer->bits_per_word);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600945 mas->cur_xfer = NULL;
946 ret = -ETIMEDOUT;
947 goto err_fifo_geni_transfer_one;
948 }
949 } else {
950 setup_gsi_xfer(xfer, mas, slv, spi);
951 if ((mas->num_xfers >= NUM_SPI_XFER) ||
952 (list_is_last(&xfer->transfer_list,
953 &spi->cur_msg->transfers))) {
954 int i;
955
956 for (i = 0 ; i < mas->num_tx_eot; i++) {
957 timeout =
958 wait_for_completion_interruptible_timeout(
959 &mas->tx_cb,
960 msecs_to_jiffies(SPI_XFER_TIMEOUT_MS));
961 if (timeout <= 0) {
962 GENI_SE_ERR(mas->ipc, true, mas->dev,
963 "Tx[%d] timeout%lu\n", i, timeout);
964 ret = -ETIMEDOUT;
965 goto err_gsi_geni_transfer_one;
966 }
967 }
968 for (i = 0 ; i < mas->num_rx_eot; i++) {
969 timeout =
970 wait_for_completion_interruptible_timeout(
971 &mas->rx_cb,
972 msecs_to_jiffies(SPI_XFER_TIMEOUT_MS));
973 if (timeout <= 0) {
974 GENI_SE_ERR(mas->ipc, true, mas->dev,
975 "Rx[%d] timeout%lu\n", i, timeout);
976 ret = -ETIMEDOUT;
977 goto err_gsi_geni_transfer_one;
978 }
979 }
980 if (mas->qn_err) {
981 ret = -EIO;
982 mas->qn_err = false;
983 goto err_gsi_geni_transfer_one;
984 }
985 }
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700986 }
Girish Mahadevan9a7c9442017-08-15 12:10:09 -0600987 return ret;
988err_gsi_geni_transfer_one:
989 dmaengine_terminate_all(mas->tx);
990 return ret;
991err_fifo_geni_transfer_one:
992 handle_fifo_timeout(mas);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -0700993 return ret;
994}
995
996static void geni_spi_handle_tx(struct spi_geni_master *mas)
997{
998 int i = 0;
999 int tx_fifo_width = (mas->tx_fifo_width >> 3);
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001000 int max_bytes = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001001 const u8 *tx_buf = NULL;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001002
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001003 if (!mas->cur_xfer)
1004 return;
1005
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001006 /*
1007 * For non-byte aligned bits-per-word values:
1008 * Assumption is that each SPI word will be accomodated in
1009 * ceil (bits_per_word / bits_per_byte)
1010 * and the next SPI word starts at the next byte.
1011 * In such cases, we can fit 1 SPI word per FIFO word so adjust the
1012 * max byte that can be sent per IRQ accordingly.
1013 */
1014 if ((mas->tx_fifo_width % mas->cur_word_len))
1015 max_bytes = (mas->tx_fifo_depth - mas->tx_wm) *
1016 ((mas->cur_word_len / BITS_PER_BYTE) + 1);
1017 else
1018 max_bytes = (mas->tx_fifo_depth - mas->tx_wm) * tx_fifo_width;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001019 tx_buf = mas->cur_xfer->tx_buf;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001020 tx_buf += (mas->cur_xfer->len - mas->tx_rem_bytes);
1021 max_bytes = min_t(int, mas->tx_rem_bytes, max_bytes);
1022 while (i < max_bytes) {
1023 int j;
1024 u32 fifo_word = 0;
1025 u8 *fifo_byte;
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001026 int bytes_per_fifo = tx_fifo_width;
1027 int bytes_to_write = 0;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001028
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001029 if ((mas->tx_fifo_width % mas->cur_word_len))
1030 bytes_per_fifo =
1031 (mas->cur_word_len / BITS_PER_BYTE) + 1;
1032 bytes_to_write = min_t(int, (max_bytes - i), bytes_per_fifo);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001033 fifo_byte = (u8 *)&fifo_word;
1034 for (j = 0; j < bytes_to_write; j++)
1035 fifo_byte[j] = tx_buf[i++];
1036 geni_write_reg(fifo_word, mas->base, SE_GENI_TX_FIFOn);
1037 /* Ensure FIFO writes are written in order */
1038 mb();
1039 }
1040 mas->tx_rem_bytes -= max_bytes;
1041 if (!mas->tx_rem_bytes) {
1042 geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG);
1043 /* Barrier here before return to prevent further ISRs */
1044 mb();
1045 }
1046}
1047
1048static void geni_spi_handle_rx(struct spi_geni_master *mas)
1049{
1050 int i = 0;
1051 int fifo_width = (mas->tx_fifo_width >> 3);
1052 u32 rx_fifo_status = geni_read_reg(mas->base, SE_GENI_RX_FIFO_STATUS);
1053 int rx_bytes = 0;
1054 int rx_wc = 0;
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001055 u8 *rx_buf = NULL;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001056
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001057 if (!mas->cur_xfer)
1058 return;
1059
1060 rx_buf = mas->cur_xfer->rx_buf;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001061 rx_wc = (rx_fifo_status & RX_FIFO_WC_MSK);
1062 if (rx_fifo_status & RX_LAST) {
1063 int rx_last_byte_valid =
1064 (rx_fifo_status & RX_LAST_BYTE_VALID_MSK)
1065 >> RX_LAST_BYTE_VALID_SHFT;
1066 if (rx_last_byte_valid && (rx_last_byte_valid < 4)) {
1067 rx_wc -= 1;
1068 rx_bytes += rx_last_byte_valid;
1069 }
1070 }
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001071 if (!(mas->tx_fifo_width % mas->cur_word_len))
1072 rx_bytes += rx_wc * fifo_width;
1073 else
1074 rx_bytes += rx_wc *
1075 ((mas->cur_word_len / BITS_PER_BYTE) + 1);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001076 rx_bytes = min_t(int, mas->rx_rem_bytes, rx_bytes);
1077 rx_buf += (mas->cur_xfer->len - mas->rx_rem_bytes);
1078 while (i < rx_bytes) {
1079 u32 fifo_word = 0;
1080 u8 *fifo_byte;
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001081 int bytes_per_fifo = fifo_width;
1082 int read_bytes = 0;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001083 int j;
1084
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001085 if ((mas->tx_fifo_width % mas->cur_word_len))
1086 bytes_per_fifo =
1087 (mas->cur_word_len / BITS_PER_BYTE) + 1;
1088 read_bytes = min_t(int, (rx_bytes - i), bytes_per_fifo);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001089 fifo_word = geni_read_reg(mas->base, SE_GENI_RX_FIFOn);
1090 fifo_byte = (u8 *)&fifo_word;
1091 for (j = 0; j < read_bytes; j++)
1092 rx_buf[i++] = fifo_byte[j];
1093 }
1094 mas->rx_rem_bytes -= rx_bytes;
1095}
1096
1097static irqreturn_t geni_spi_irq(int irq, void *dev)
1098{
1099 struct spi_geni_master *mas = dev;
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001100 u32 m_irq = 0;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001101
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001102 if (pm_runtime_status_suspended(dev)) {
1103 GENI_SE_DBG(mas->ipc, false, mas->dev,
1104 "%s: device is suspended\n", __func__);
1105 goto exit_geni_spi_irq;
1106 }
1107 m_irq = geni_read_reg(mas->base, SE_GENI_M_IRQ_STATUS);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001108 if ((m_irq & M_RX_FIFO_WATERMARK_EN) || (m_irq & M_RX_FIFO_LAST_EN))
1109 geni_spi_handle_rx(mas);
1110
1111 if ((m_irq & M_TX_FIFO_WATERMARK_EN))
1112 geni_spi_handle_tx(mas);
1113
1114 if ((m_irq & M_CMD_DONE_EN) || (m_irq & M_CMD_CANCEL_EN) ||
1115 (m_irq & M_CMD_ABORT_EN)) {
1116 complete(&mas->xfer_done);
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001117 /*
1118 * If this happens, then a CMD_DONE came before all the buffer
1119 * bytes were sent out. This is unusual, log this condition and
1120 * disable the WM interrupt to prevent the system from stalling
1121 * due an interrupt storm.
1122 * If this happens when all Rx bytes haven't been received, log
1123 * the condition.
1124 */
1125 if (mas->tx_rem_bytes) {
1126 geni_write_reg(0, mas->base, SE_GENI_TX_WATERMARK_REG);
1127 GENI_SE_DBG(mas->ipc, false, mas->dev,
1128 "%s:Premature Done.tx_rem%d bpw%d\n",
1129 __func__, mas->tx_rem_bytes, mas->cur_word_len);
1130 }
1131 if (mas->rx_rem_bytes)
1132 GENI_SE_DBG(mas->ipc, false, mas->dev,
1133 "%s:Premature Done.rx_rem%d bpw%d\n",
1134 __func__, mas->rx_rem_bytes, mas->cur_word_len);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001135 }
Girish Mahadevan1fc27702017-08-31 12:55:23 -06001136exit_geni_spi_irq:
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001137 geni_write_reg(m_irq, mas->base, SE_GENI_M_IRQ_CLEAR);
1138 return IRQ_HANDLED;
1139}
1140
1141static int spi_geni_probe(struct platform_device *pdev)
1142{
1143 int ret;
1144 struct spi_master *spi;
1145 struct spi_geni_master *geni_mas;
1146 struct se_geni_rsc *rsc;
1147 struct resource *res;
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -06001148 struct platform_device *wrapper_pdev;
1149 struct device_node *wrapper_ph_node;
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001150
1151 spi = spi_alloc_master(&pdev->dev, sizeof(struct spi_geni_master));
1152 if (!spi) {
1153 ret = -ENOMEM;
1154 dev_err(&pdev->dev, "Failed to alloc spi struct\n");
1155 goto spi_geni_probe_err;
1156 }
1157
1158 platform_set_drvdata(pdev, spi);
1159 geni_mas = spi_master_get_devdata(spi);
1160 rsc = &geni_mas->spi_rsc;
1161 geni_mas->dev = &pdev->dev;
1162 spi->dev.of_node = pdev->dev.of_node;
Karthikeyan Ramasubramanian0d578b72017-04-26 10:44:02 -06001163 wrapper_ph_node = of_parse_phandle(pdev->dev.of_node,
1164 "qcom,wrapper-core", 0);
1165 if (IS_ERR_OR_NULL(wrapper_ph_node)) {
1166 ret = PTR_ERR(wrapper_ph_node);
1167 dev_err(&pdev->dev, "No wrapper core defined\n");
1168 goto spi_geni_probe_err;
1169 }
1170 wrapper_pdev = of_find_device_by_node(wrapper_ph_node);
1171 of_node_put(wrapper_ph_node);
1172 if (IS_ERR_OR_NULL(wrapper_pdev)) {
1173 ret = PTR_ERR(wrapper_pdev);
1174 dev_err(&pdev->dev, "Cannot retrieve wrapper device\n");
1175 goto spi_geni_probe_err;
1176 }
1177 geni_mas->wrapper_dev = &wrapper_pdev->dev;
1178 geni_mas->spi_rsc.wrapper_dev = &wrapper_pdev->dev;
1179 ret = geni_se_resources_init(rsc, SPI_CORE2X_VOTE,
1180 (DEFAULT_SE_CLK * DEFAULT_BUS_WIDTH));
1181 if (ret) {
1182 dev_err(&pdev->dev, "Error geni_se_resources_init\n");
1183 goto spi_geni_probe_err;
1184 }
1185
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001186 rsc->geni_pinctrl = devm_pinctrl_get(&pdev->dev);
1187 if (IS_ERR_OR_NULL(rsc->geni_pinctrl)) {
1188 dev_err(&pdev->dev, "No pinctrl config specified!\n");
1189 ret = PTR_ERR(rsc->geni_pinctrl);
1190 goto spi_geni_probe_err;
1191 }
1192
1193 rsc->geni_gpio_active = pinctrl_lookup_state(rsc->geni_pinctrl,
1194 PINCTRL_DEFAULT);
1195 if (IS_ERR_OR_NULL(rsc->geni_gpio_active)) {
1196 dev_err(&pdev->dev, "No default config specified!\n");
1197 ret = PTR_ERR(rsc->geni_gpio_active);
1198 goto spi_geni_probe_err;
1199 }
1200
1201 rsc->geni_gpio_sleep = pinctrl_lookup_state(rsc->geni_pinctrl,
1202 PINCTRL_SLEEP);
1203 if (IS_ERR_OR_NULL(rsc->geni_gpio_sleep)) {
1204 dev_err(&pdev->dev, "No sleep config specified!\n");
1205 ret = PTR_ERR(rsc->geni_gpio_sleep);
1206 goto spi_geni_probe_err;
1207 }
1208
1209 rsc->se_clk = devm_clk_get(&pdev->dev, "se-clk");
1210 if (IS_ERR(rsc->se_clk)) {
1211 ret = PTR_ERR(rsc->se_clk);
1212 dev_err(&pdev->dev, "Err getting SE Core clk %d\n", ret);
1213 goto spi_geni_probe_err;
1214 }
1215
1216 rsc->m_ahb_clk = devm_clk_get(&pdev->dev, "m-ahb");
1217 if (IS_ERR(rsc->m_ahb_clk)) {
1218 ret = PTR_ERR(rsc->m_ahb_clk);
1219 dev_err(&pdev->dev, "Err getting M AHB clk %d\n", ret);
1220 goto spi_geni_probe_err;
1221 }
1222
1223 rsc->s_ahb_clk = devm_clk_get(&pdev->dev, "s-ahb");
1224 if (IS_ERR(rsc->s_ahb_clk)) {
1225 ret = PTR_ERR(rsc->s_ahb_clk);
1226 dev_err(&pdev->dev, "Err getting S AHB clk %d\n", ret);
1227 goto spi_geni_probe_err;
1228 }
1229
1230 if (of_property_read_u32(pdev->dev.of_node, "spi-max-frequency",
1231 &spi->max_speed_hz)) {
1232 dev_err(&pdev->dev, "Max frequency not specified.\n");
1233 ret = -ENXIO;
1234 goto spi_geni_probe_err;
1235 }
1236
1237 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "se_phys");
1238 if (!res) {
1239 ret = -ENXIO;
1240 dev_err(&pdev->dev, "Err getting IO region\n");
1241 goto spi_geni_probe_err;
1242 }
1243
1244 geni_mas->phys_addr = res->start;
1245 geni_mas->size = resource_size(res);
1246 geni_mas->base = devm_ioremap(&pdev->dev, res->start,
1247 resource_size(res));
1248 if (!geni_mas->base) {
1249 ret = -ENOMEM;
1250 dev_err(&pdev->dev, "Err IO mapping iomem\n");
1251 goto spi_geni_probe_err;
1252 }
1253
1254 geni_mas->irq = platform_get_irq(pdev, 0);
1255 if (geni_mas->irq < 0) {
1256 dev_err(&pdev->dev, "Err getting IRQ\n");
1257 ret = geni_mas->irq;
1258 goto spi_geni_probe_unmap;
1259 }
1260 ret = devm_request_irq(&pdev->dev, geni_mas->irq, geni_spi_irq,
1261 IRQF_TRIGGER_HIGH, "spi_geni", geni_mas);
1262 if (ret) {
1263 dev_err(&pdev->dev, "Request_irq failed:%d: err:%d\n",
1264 geni_mas->irq, ret);
1265 goto spi_geni_probe_unmap;
1266 }
1267
1268 spi->mode_bits = (SPI_CPOL | SPI_CPHA | SPI_LOOP | SPI_CS_HIGH);
1269 spi->bits_per_word_mask = SPI_BPW_RANGE_MASK(4, 32);
1270 spi->num_chipselect = SPI_NUM_CHIPSELECT;
1271 spi->prepare_transfer_hardware = spi_geni_prepare_transfer_hardware;
1272 spi->prepare_message = spi_geni_prepare_message;
1273 spi->unprepare_message = spi_geni_unprepare_message;
1274 spi->transfer_one = spi_geni_transfer_one;
1275 spi->unprepare_transfer_hardware
1276 = spi_geni_unprepare_transfer_hardware;
1277 spi->auto_runtime_pm = false;
1278
1279 init_completion(&geni_mas->xfer_done);
Girish Mahadevan9a7c9442017-08-15 12:10:09 -06001280 init_completion(&geni_mas->tx_cb);
1281 init_completion(&geni_mas->rx_cb);
Girish Mahadevan2ef85af2017-02-14 14:42:22 -07001282 pm_runtime_enable(&pdev->dev);
1283 ret = spi_register_master(spi);
1284 if (ret) {
1285 dev_err(&pdev->dev, "Failed to register SPI master\n");
1286 goto spi_geni_probe_unmap;
1287 }
1288 return ret;
1289spi_geni_probe_unmap:
1290 devm_iounmap(&pdev->dev, geni_mas->base);
1291spi_geni_probe_err:
1292 spi_master_put(spi);
1293 return ret;
1294}
1295
1296static int spi_geni_remove(struct platform_device *pdev)
1297{
1298 struct spi_master *master = platform_get_drvdata(pdev);
1299 struct spi_geni_master *geni_mas = spi_master_get_devdata(master);
1300
1301 spi_unregister_master(master);
1302 se_geni_resources_off(&geni_mas->spi_rsc);
1303 pm_runtime_put_noidle(&pdev->dev);
1304 pm_runtime_disable(&pdev->dev);
1305 return 0;
1306}
1307
1308#ifdef CONFIG_PM
1309static int spi_geni_runtime_suspend(struct device *dev)
1310{
1311 int ret = 0;
1312 struct spi_master *spi = get_spi_master(dev);
1313 struct spi_geni_master *geni_mas = spi_master_get_devdata(spi);
1314
1315 ret = se_geni_resources_off(&geni_mas->spi_rsc);
1316 return ret;
1317}
1318
1319static int spi_geni_runtime_resume(struct device *dev)
1320{
1321 int ret = 0;
1322 struct spi_master *spi = get_spi_master(dev);
1323 struct spi_geni_master *geni_mas = spi_master_get_devdata(spi);
1324
1325 ret = se_geni_resources_on(&geni_mas->spi_rsc);
1326 return ret;
1327}
1328
1329static int spi_geni_resume(struct device *dev)
1330{
1331 return 0;
1332}
1333
1334static int spi_geni_suspend(struct device *dev)
1335{
1336 if (!pm_runtime_status_suspended(dev))
1337 return -EBUSY;
1338 return 0;
1339}
1340#else
1341static int spi_geni_runtime_suspend(struct device *dev)
1342{
1343 return 0;
1344}
1345
1346static int spi_geni_runtime_resume(struct device *dev)
1347{
1348 return 0;
1349}
1350
1351static int spi_geni_resume(struct device *dev)
1352{
1353 return 0;
1354}
1355
1356static int spi_geni_suspend(struct device *dev)
1357{
1358 return 0;
1359}
1360#endif
1361
1362static const struct dev_pm_ops spi_geni_pm_ops = {
1363 SET_RUNTIME_PM_OPS(spi_geni_runtime_suspend,
1364 spi_geni_runtime_resume, NULL)
1365 SET_SYSTEM_SLEEP_PM_OPS(spi_geni_suspend, spi_geni_resume)
1366};
1367
1368static const struct of_device_id spi_geni_dt_match[] = {
1369 { .compatible = "qcom,spi-geni" },
1370 {}
1371};
1372
1373static struct platform_driver spi_geni_driver = {
1374 .probe = spi_geni_probe,
1375 .remove = spi_geni_remove,
1376 .driver = {
1377 .name = "spi_geni",
1378 .pm = &spi_geni_pm_ops,
1379 .of_match_table = spi_geni_dt_match,
1380 },
1381};
1382module_platform_driver(spi_geni_driver);
1383
1384MODULE_LICENSE("GPL v2");
1385MODULE_ALIAS("platform:spi_geni");