Naresh Kumar Inna | a3667aa | 2012-11-15 22:41:18 +0530 | [diff] [blame] | 1 | /* |
| 2 | * This file is part of the Chelsio FCoE driver for Linux. |
| 3 | * |
| 4 | * Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved. |
| 5 | * |
| 6 | * This software is available to you under a choice of one of two |
| 7 | * licenses. You may choose to be licensed under the terms of the GNU |
| 8 | * General Public License (GPL) Version 2, available from the file |
| 9 | * COPYING in the main directory of this source tree, or the |
| 10 | * OpenIB.org BSD license below: |
| 11 | * |
| 12 | * Redistribution and use in source and binary forms, with or |
| 13 | * without modification, are permitted provided that the following |
| 14 | * conditions are met: |
| 15 | * |
| 16 | * - Redistributions of source code must retain the above |
| 17 | * copyright notice, this list of conditions and the following |
| 18 | * disclaimer. |
| 19 | * |
| 20 | * - Redistributions in binary form must reproduce the above |
| 21 | * copyright notice, this list of conditions and the following |
| 22 | * disclaimer in the documentation and/or other materials |
| 23 | * provided with the distribution. |
| 24 | * |
| 25 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| 26 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| 27 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| 28 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| 29 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| 30 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| 31 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| 32 | * SOFTWARE. |
| 33 | */ |
| 34 | |
| 35 | #ifndef __CSIO_WR_H__ |
| 36 | #define __CSIO_WR_H__ |
| 37 | |
| 38 | #include <linux/cache.h> |
| 39 | |
| 40 | #include "csio_defs.h" |
| 41 | #include "t4fw_api.h" |
| 42 | #include "t4fw_api_stor.h" |
| 43 | |
| 44 | /* |
| 45 | * SGE register field values. |
| 46 | */ |
| 47 | #define X_INGPCIEBOUNDARY_32B 0 |
| 48 | #define X_INGPCIEBOUNDARY_64B 1 |
| 49 | #define X_INGPCIEBOUNDARY_128B 2 |
| 50 | #define X_INGPCIEBOUNDARY_256B 3 |
| 51 | #define X_INGPCIEBOUNDARY_512B 4 |
| 52 | #define X_INGPCIEBOUNDARY_1024B 5 |
| 53 | #define X_INGPCIEBOUNDARY_2048B 6 |
| 54 | #define X_INGPCIEBOUNDARY_4096B 7 |
| 55 | |
| 56 | /* GTS register */ |
| 57 | #define X_TIMERREG_COUNTER0 0 |
| 58 | #define X_TIMERREG_COUNTER1 1 |
| 59 | #define X_TIMERREG_COUNTER2 2 |
| 60 | #define X_TIMERREG_COUNTER3 3 |
| 61 | #define X_TIMERREG_COUNTER4 4 |
| 62 | #define X_TIMERREG_COUNTER5 5 |
| 63 | #define X_TIMERREG_RESTART_COUNTER 6 |
| 64 | #define X_TIMERREG_UPDATE_CIDX 7 |
| 65 | |
| 66 | /* |
| 67 | * Egress Context field values |
| 68 | */ |
| 69 | #define X_FETCHBURSTMIN_16B 0 |
| 70 | #define X_FETCHBURSTMIN_32B 1 |
| 71 | #define X_FETCHBURSTMIN_64B 2 |
| 72 | #define X_FETCHBURSTMIN_128B 3 |
| 73 | |
| 74 | #define X_FETCHBURSTMAX_64B 0 |
| 75 | #define X_FETCHBURSTMAX_128B 1 |
| 76 | #define X_FETCHBURSTMAX_256B 2 |
| 77 | #define X_FETCHBURSTMAX_512B 3 |
| 78 | |
| 79 | #define X_HOSTFCMODE_NONE 0 |
| 80 | #define X_HOSTFCMODE_INGRESS_QUEUE 1 |
| 81 | #define X_HOSTFCMODE_STATUS_PAGE 2 |
| 82 | #define X_HOSTFCMODE_BOTH 3 |
| 83 | |
| 84 | /* |
| 85 | * Ingress Context field values |
| 86 | */ |
| 87 | #define X_UPDATESCHEDULING_TIMER 0 |
| 88 | #define X_UPDATESCHEDULING_COUNTER_OPTTIMER 1 |
| 89 | |
| 90 | #define X_UPDATEDELIVERY_NONE 0 |
| 91 | #define X_UPDATEDELIVERY_INTERRUPT 1 |
| 92 | #define X_UPDATEDELIVERY_STATUS_PAGE 2 |
| 93 | #define X_UPDATEDELIVERY_BOTH 3 |
| 94 | |
| 95 | #define X_INTERRUPTDESTINATION_PCIE 0 |
| 96 | #define X_INTERRUPTDESTINATION_IQ 1 |
| 97 | |
| 98 | #define X_RSPD_TYPE_FLBUF 0 |
| 99 | #define X_RSPD_TYPE_CPL 1 |
| 100 | #define X_RSPD_TYPE_INTR 2 |
| 101 | |
| 102 | /* WR status is at the same position as retval in a CMD header */ |
| 103 | #define csio_wr_status(_wr) \ |
| 104 | (FW_CMD_RETVAL_GET(ntohl(((struct fw_cmd_hdr *)(_wr))->lo))) |
| 105 | |
| 106 | struct csio_hw; |
| 107 | |
| 108 | extern int csio_intr_coalesce_cnt; |
| 109 | extern int csio_intr_coalesce_time; |
| 110 | |
| 111 | /* Ingress queue params */ |
| 112 | struct csio_iq_params { |
| 113 | |
| 114 | uint8_t iq_start:1; |
| 115 | uint8_t iq_stop:1; |
| 116 | uint8_t pfn:3; |
| 117 | |
| 118 | uint8_t vfn; |
| 119 | |
| 120 | uint16_t physiqid; |
| 121 | uint16_t iqid; |
| 122 | |
| 123 | uint16_t fl0id; |
| 124 | uint16_t fl1id; |
| 125 | |
| 126 | uint8_t viid; |
| 127 | |
| 128 | uint8_t type; |
| 129 | uint8_t iqasynch; |
| 130 | uint8_t reserved4; |
| 131 | |
| 132 | uint8_t iqandst; |
| 133 | uint8_t iqanus; |
| 134 | uint8_t iqanud; |
| 135 | |
| 136 | uint16_t iqandstindex; |
| 137 | |
| 138 | uint8_t iqdroprss; |
| 139 | uint8_t iqpciech; |
| 140 | uint8_t iqdcaen; |
| 141 | |
| 142 | uint8_t iqdcacpu; |
| 143 | uint8_t iqintcntthresh; |
| 144 | uint8_t iqo; |
| 145 | |
| 146 | uint8_t iqcprio; |
| 147 | uint8_t iqesize; |
| 148 | |
| 149 | uint16_t iqsize; |
| 150 | |
| 151 | uint64_t iqaddr; |
| 152 | |
| 153 | uint8_t iqflintiqhsen; |
| 154 | uint8_t reserved5; |
| 155 | uint8_t iqflintcongen; |
| 156 | uint8_t iqflintcngchmap; |
| 157 | |
| 158 | uint32_t reserved6; |
| 159 | |
| 160 | uint8_t fl0hostfcmode; |
| 161 | uint8_t fl0cprio; |
| 162 | uint8_t fl0paden; |
| 163 | uint8_t fl0packen; |
| 164 | uint8_t fl0congen; |
| 165 | uint8_t fl0dcaen; |
| 166 | |
| 167 | uint8_t fl0dcacpu; |
| 168 | uint8_t fl0fbmin; |
| 169 | |
| 170 | uint8_t fl0fbmax; |
| 171 | uint8_t fl0cidxfthresho; |
| 172 | uint8_t fl0cidxfthresh; |
| 173 | |
| 174 | uint16_t fl0size; |
| 175 | |
| 176 | uint64_t fl0addr; |
| 177 | |
| 178 | uint64_t reserved7; |
| 179 | |
| 180 | uint8_t fl1hostfcmode; |
| 181 | uint8_t fl1cprio; |
| 182 | uint8_t fl1paden; |
| 183 | uint8_t fl1packen; |
| 184 | uint8_t fl1congen; |
| 185 | uint8_t fl1dcaen; |
| 186 | |
| 187 | uint8_t fl1dcacpu; |
| 188 | uint8_t fl1fbmin; |
| 189 | |
| 190 | uint8_t fl1fbmax; |
| 191 | uint8_t fl1cidxfthresho; |
| 192 | uint8_t fl1cidxfthresh; |
| 193 | |
| 194 | uint16_t fl1size; |
| 195 | |
| 196 | uint64_t fl1addr; |
| 197 | }; |
| 198 | |
| 199 | /* Egress queue params */ |
| 200 | struct csio_eq_params { |
| 201 | |
| 202 | uint8_t pfn; |
| 203 | uint8_t vfn; |
| 204 | |
| 205 | uint8_t eqstart:1; |
| 206 | uint8_t eqstop:1; |
| 207 | |
| 208 | uint16_t physeqid; |
| 209 | uint32_t eqid; |
| 210 | |
| 211 | uint8_t hostfcmode:2; |
| 212 | uint8_t cprio:1; |
| 213 | uint8_t pciechn:3; |
| 214 | |
| 215 | uint16_t iqid; |
| 216 | |
| 217 | uint8_t dcaen:1; |
| 218 | uint8_t dcacpu:5; |
| 219 | |
| 220 | uint8_t fbmin:3; |
| 221 | uint8_t fbmax:3; |
| 222 | |
| 223 | uint8_t cidxfthresho:1; |
| 224 | uint8_t cidxfthresh:3; |
| 225 | |
| 226 | uint16_t eqsize; |
| 227 | |
| 228 | uint64_t eqaddr; |
| 229 | }; |
| 230 | |
| 231 | struct csio_dma_buf { |
| 232 | struct list_head list; |
| 233 | void *vaddr; /* Virtual address */ |
| 234 | dma_addr_t paddr; /* Physical address */ |
| 235 | uint32_t len; /* Buffer size */ |
| 236 | }; |
| 237 | |
| 238 | /* Generic I/O request structure */ |
| 239 | struct csio_ioreq { |
| 240 | struct csio_sm sm; /* SM, List |
| 241 | * should be the first member |
| 242 | */ |
| 243 | int iq_idx; /* Ingress queue index */ |
| 244 | int eq_idx; /* Egress queue index */ |
| 245 | uint32_t nsge; /* Number of SG elements */ |
| 246 | uint32_t tmo; /* Driver timeout */ |
| 247 | uint32_t datadir; /* Data direction */ |
| 248 | struct csio_dma_buf dma_buf; /* Req/resp DMA buffers */ |
| 249 | uint16_t wr_status; /* WR completion status */ |
| 250 | int16_t drv_status; /* Driver internal status */ |
| 251 | struct csio_lnode *lnode; /* Owner lnode */ |
| 252 | struct csio_rnode *rnode; /* Src/destination rnode */ |
| 253 | void (*io_cbfn) (struct csio_hw *, struct csio_ioreq *); |
| 254 | /* completion callback */ |
| 255 | void *scratch1; /* Scratch area 1. |
| 256 | */ |
| 257 | void *scratch2; /* Scratch area 2. */ |
| 258 | struct list_head gen_list; /* Any list associated with |
| 259 | * this ioreq. |
| 260 | */ |
| 261 | uint64_t fw_handle; /* Unique handle passed |
| 262 | * to FW |
| 263 | */ |
| 264 | uint8_t dcopy; /* Data copy required */ |
| 265 | uint8_t reserved1; |
| 266 | uint16_t reserved2; |
| 267 | struct completion cmplobj; /* ioreq completion object */ |
| 268 | } ____cacheline_aligned_in_smp; |
| 269 | |
| 270 | /* |
| 271 | * Egress status page for egress cidx updates |
| 272 | */ |
| 273 | struct csio_qstatus_page { |
| 274 | __be32 qid; |
| 275 | __be16 cidx; |
| 276 | __be16 pidx; |
| 277 | }; |
| 278 | |
| 279 | |
| 280 | enum { |
| 281 | CSIO_MAX_FLBUF_PER_IQWR = 4, |
| 282 | CSIO_QCREDIT_SZ = 64, /* pidx/cidx increments |
| 283 | * in bytes |
| 284 | */ |
| 285 | CSIO_MAX_QID = 0xFFFF, |
| 286 | CSIO_MAX_IQ = 128, |
| 287 | |
| 288 | CSIO_SGE_NTIMERS = 6, |
| 289 | CSIO_SGE_NCOUNTERS = 4, |
| 290 | CSIO_SGE_FL_SIZE_REGS = 16, |
| 291 | }; |
| 292 | |
| 293 | /* Defines for type */ |
| 294 | enum { |
| 295 | CSIO_EGRESS = 1, |
| 296 | CSIO_INGRESS = 2, |
| 297 | CSIO_FREELIST = 3, |
| 298 | }; |
| 299 | |
| 300 | /* |
| 301 | * Structure for footer (last 2 flits) of Ingress Queue Entry. |
| 302 | */ |
| 303 | struct csio_iqwr_footer { |
| 304 | __be32 hdrbuflen_pidx; |
| 305 | __be32 pldbuflen_qid; |
| 306 | union { |
| 307 | u8 type_gen; |
| 308 | __be64 last_flit; |
| 309 | } u; |
| 310 | }; |
| 311 | |
| 312 | #define IQWRF_NEWBUF (1 << 31) |
| 313 | #define IQWRF_LEN_GET(x) (((x) >> 0) & 0x7fffffffU) |
| 314 | #define IQWRF_GEN_SHIFT 7 |
| 315 | #define IQWRF_TYPE_GET(x) (((x) >> 4) & 0x3U) |
| 316 | |
| 317 | |
| 318 | /* |
| 319 | * WR pair: |
| 320 | * ======== |
| 321 | * A WR can start towards the end of a queue, and then continue at the |
| 322 | * beginning, since the queue is considered to be circular. This will |
| 323 | * require a pair of address/len to be passed back to the caller - |
| 324 | * hence the Work request pair structure. |
| 325 | */ |
| 326 | struct csio_wr_pair { |
| 327 | void *addr1; |
| 328 | uint32_t size1; |
| 329 | void *addr2; |
| 330 | uint32_t size2; |
| 331 | }; |
| 332 | |
| 333 | /* |
| 334 | * The following structure is used by ingress processing to return the |
| 335 | * free list buffers to consumers. |
| 336 | */ |
| 337 | struct csio_fl_dma_buf { |
| 338 | struct csio_dma_buf flbufs[CSIO_MAX_FLBUF_PER_IQWR]; |
| 339 | /* Freelist DMA buffers */ |
| 340 | int offset; /* Offset within the |
| 341 | * first FL buf. |
| 342 | */ |
| 343 | uint32_t totlen; /* Total length */ |
| 344 | uint8_t defer_free; /* Free of buffer can |
| 345 | * deferred |
| 346 | */ |
| 347 | }; |
| 348 | |
| 349 | /* Data-types */ |
| 350 | typedef void (*iq_handler_t)(struct csio_hw *, void *, uint32_t, |
| 351 | struct csio_fl_dma_buf *, void *); |
| 352 | |
| 353 | struct csio_iq { |
| 354 | uint16_t iqid; /* Queue ID */ |
| 355 | uint16_t physiqid; /* Physical Queue ID */ |
| 356 | uint16_t genbit; /* Generation bit, |
| 357 | * initially set to 1 |
| 358 | */ |
| 359 | int flq_idx; /* Freelist queue index */ |
| 360 | iq_handler_t iq_intx_handler; /* IQ INTx handler routine */ |
| 361 | }; |
| 362 | |
| 363 | struct csio_eq { |
| 364 | uint16_t eqid; /* Qid */ |
| 365 | uint16_t physeqid; /* Physical Queue ID */ |
| 366 | uint8_t wrap[512]; /* Temp area for q-wrap around*/ |
| 367 | }; |
| 368 | |
| 369 | struct csio_fl { |
| 370 | uint16_t flid; /* Qid */ |
| 371 | uint16_t packen; /* Packing enabled? */ |
| 372 | int offset; /* Offset within FL buf */ |
| 373 | int sreg; /* Size register */ |
| 374 | struct csio_dma_buf *bufs; /* Free list buffer ptr array |
| 375 | * indexed using flq->cidx/pidx |
| 376 | */ |
| 377 | }; |
| 378 | |
| 379 | struct csio_qstats { |
| 380 | uint32_t n_tot_reqs; /* Total no. of Requests */ |
| 381 | uint32_t n_tot_rsps; /* Total no. of responses */ |
| 382 | uint32_t n_qwrap; /* Queue wraps */ |
| 383 | uint32_t n_eq_wr_split; /* Number of split EQ WRs */ |
| 384 | uint32_t n_qentry; /* Queue entry */ |
| 385 | uint32_t n_qempty; /* Queue empty */ |
| 386 | uint32_t n_qfull; /* Queue fulls */ |
| 387 | uint32_t n_rsp_unknown; /* Unknown response type */ |
| 388 | uint32_t n_stray_comp; /* Stray completion intr */ |
| 389 | uint32_t n_flq_refill; /* Number of FL refills */ |
| 390 | }; |
| 391 | |
| 392 | /* Queue metadata */ |
| 393 | struct csio_q { |
| 394 | uint16_t type; /* Type: Ingress/Egress/FL */ |
| 395 | uint16_t pidx; /* producer index */ |
| 396 | uint16_t cidx; /* consumer index */ |
| 397 | uint16_t inc_idx; /* Incremental index */ |
| 398 | uint32_t wr_sz; /* Size of all WRs in this q |
| 399 | * if fixed |
| 400 | */ |
| 401 | void *vstart; /* Base virtual address |
| 402 | * of queue |
| 403 | */ |
| 404 | void *vwrap; /* Virtual end address to |
| 405 | * wrap around at |
| 406 | */ |
| 407 | uint32_t credits; /* Size of queue in credits */ |
| 408 | void *owner; /* Owner */ |
| 409 | union { /* Queue contexts */ |
| 410 | struct csio_iq iq; |
| 411 | struct csio_eq eq; |
| 412 | struct csio_fl fl; |
| 413 | } un; |
| 414 | |
| 415 | dma_addr_t pstart; /* Base physical address of |
| 416 | * queue |
| 417 | */ |
| 418 | uint32_t portid; /* PCIE Channel */ |
| 419 | uint32_t size; /* Size of queue in bytes */ |
| 420 | struct csio_qstats stats; /* Statistics */ |
| 421 | } ____cacheline_aligned_in_smp; |
| 422 | |
| 423 | struct csio_sge { |
| 424 | uint32_t csio_fl_align; /* Calculated and cached |
| 425 | * for fast path |
| 426 | */ |
| 427 | uint32_t sge_control; /* padding, boundaries, |
| 428 | * lengths, etc. |
| 429 | */ |
| 430 | uint32_t sge_host_page_size; /* Host page size */ |
| 431 | uint32_t sge_fl_buf_size[CSIO_SGE_FL_SIZE_REGS]; |
| 432 | /* free list buffer sizes */ |
| 433 | uint16_t timer_val[CSIO_SGE_NTIMERS]; |
| 434 | uint8_t counter_val[CSIO_SGE_NCOUNTERS]; |
| 435 | }; |
| 436 | |
| 437 | /* Work request module */ |
| 438 | struct csio_wrm { |
| 439 | int num_q; /* Number of queues */ |
| 440 | struct csio_q **q_arr; /* Array of queue pointers |
| 441 | * allocated dynamically |
| 442 | * based on configured values |
| 443 | */ |
| 444 | uint32_t fw_iq_start; /* Start ID of IQ for this fn*/ |
| 445 | uint32_t fw_eq_start; /* Start ID of EQ for this fn*/ |
| 446 | struct csio_q *intr_map[CSIO_MAX_IQ]; |
| 447 | /* IQ-id to IQ map table. */ |
| 448 | int free_qidx; /* queue idx of free queue */ |
| 449 | struct csio_sge sge; /* SGE params */ |
| 450 | }; |
| 451 | |
| 452 | #define csio_get_q(__hw, __idx) ((__hw)->wrm.q_arr[__idx]) |
| 453 | #define csio_q_type(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->type) |
| 454 | #define csio_q_pidx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->pidx) |
| 455 | #define csio_q_cidx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->cidx) |
| 456 | #define csio_q_inc_idx(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->inc_idx) |
| 457 | #define csio_q_vstart(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->vstart) |
| 458 | #define csio_q_pstart(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->pstart) |
| 459 | #define csio_q_size(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->size) |
| 460 | #define csio_q_credits(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->credits) |
| 461 | #define csio_q_portid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->portid) |
| 462 | #define csio_q_wr_sz(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->wr_sz) |
| 463 | #define csio_q_iqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.iq.iqid) |
| 464 | #define csio_q_physiqid(__hw, __idx) \ |
| 465 | ((__hw)->wrm.q_arr[(__idx)]->un.iq.physiqid) |
| 466 | #define csio_q_iq_flq_idx(__hw, __idx) \ |
| 467 | ((__hw)->wrm.q_arr[(__idx)]->un.iq.flq_idx) |
| 468 | #define csio_q_eqid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.eqid) |
| 469 | #define csio_q_flid(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.fl.flid) |
| 470 | |
| 471 | #define csio_q_physeqid(__hw, __idx) \ |
| 472 | ((__hw)->wrm.q_arr[(__idx)]->un.eq.physeqid) |
| 473 | #define csio_iq_has_fl(__iq) ((__iq)->un.iq.flq_idx != -1) |
| 474 | |
| 475 | #define csio_q_iq_to_flid(__hw, __iq_idx) \ |
| 476 | csio_q_flid((__hw), (__hw)->wrm.q_arr[(__iq_qidx)]->un.iq.flq_idx) |
| 477 | #define csio_q_set_intr_map(__hw, __iq_idx, __rel_iq_id) \ |
| 478 | (__hw)->wrm.intr_map[__rel_iq_id] = csio_get_q(__hw, __iq_idx) |
| 479 | #define csio_q_eq_wrap(__hw, __idx) ((__hw)->wrm.q_arr[(__idx)]->un.eq.wrap) |
| 480 | |
| 481 | struct csio_mb; |
| 482 | |
| 483 | int csio_wr_alloc_q(struct csio_hw *, uint32_t, uint32_t, |
| 484 | uint16_t, void *, uint32_t, int, iq_handler_t); |
| 485 | int csio_wr_iq_create(struct csio_hw *, void *, int, |
| 486 | uint32_t, uint8_t, bool, |
| 487 | void (*)(struct csio_hw *, struct csio_mb *)); |
| 488 | int csio_wr_eq_create(struct csio_hw *, void *, int, int, uint8_t, |
| 489 | void (*)(struct csio_hw *, struct csio_mb *)); |
| 490 | int csio_wr_destroy_queues(struct csio_hw *, bool cmd); |
| 491 | |
| 492 | |
| 493 | int csio_wr_get(struct csio_hw *, int, uint32_t, |
| 494 | struct csio_wr_pair *); |
| 495 | void csio_wr_copy_to_wrp(void *, struct csio_wr_pair *, uint32_t, uint32_t); |
| 496 | int csio_wr_issue(struct csio_hw *, int, bool); |
| 497 | int csio_wr_process_iq(struct csio_hw *, struct csio_q *, |
| 498 | void (*)(struct csio_hw *, void *, |
| 499 | uint32_t, struct csio_fl_dma_buf *, |
| 500 | void *), |
| 501 | void *); |
| 502 | int csio_wr_process_iq_idx(struct csio_hw *, int, |
| 503 | void (*)(struct csio_hw *, void *, |
| 504 | uint32_t, struct csio_fl_dma_buf *, |
| 505 | void *), |
| 506 | void *); |
| 507 | |
| 508 | void csio_wr_sge_init(struct csio_hw *); |
| 509 | int csio_wrm_init(struct csio_wrm *, struct csio_hw *); |
| 510 | void csio_wrm_exit(struct csio_wrm *, struct csio_hw *); |
| 511 | |
| 512 | #endif /* ifndef __CSIO_WR_H__ */ |