Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 1 | /* QLogic qede NIC Driver |
| 2 | * Copyright (c) 2015 QLogic Corporation |
| 3 | * |
| 4 | * This software is available under the terms of the GNU General Public License |
| 5 | * (GPL) Version 2, available from the file COPYING in the main directory of |
| 6 | * this source tree. |
| 7 | */ |
| 8 | |
| 9 | #ifndef _QEDE_H_ |
| 10 | #define _QEDE_H_ |
| 11 | #include <linux/compiler.h> |
| 12 | #include <linux/version.h> |
| 13 | #include <linux/workqueue.h> |
| 14 | #include <linux/netdevice.h> |
| 15 | #include <linux/interrupt.h> |
| 16 | #include <linux/bitmap.h> |
| 17 | #include <linux/kernel.h> |
| 18 | #include <linux/mutex.h> |
| 19 | #include <linux/io.h> |
| 20 | #include <linux/qed/common_hsi.h> |
| 21 | #include <linux/qed/eth_common.h> |
| 22 | #include <linux/qed/qed_if.h> |
| 23 | #include <linux/qed/qed_chain.h> |
| 24 | #include <linux/qed/qed_eth_if.h> |
| 25 | |
| 26 | #define QEDE_MAJOR_VERSION 8 |
Yuval Mintz | d4ee528 | 2016-02-21 11:40:11 +0200 | [diff] [blame] | 27 | #define QEDE_MINOR_VERSION 7 |
Yuval Mintz | 7c2d7d7 | 2016-04-10 12:43:02 +0300 | [diff] [blame] | 28 | #define QEDE_REVISION_VERSION 1 |
| 29 | #define QEDE_ENGINEERING_VERSION 20 |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 30 | #define DRV_MODULE_VERSION __stringify(QEDE_MAJOR_VERSION) "." \ |
| 31 | __stringify(QEDE_MINOR_VERSION) "." \ |
| 32 | __stringify(QEDE_REVISION_VERSION) "." \ |
| 33 | __stringify(QEDE_ENGINEERING_VERSION) |
| 34 | |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 35 | #define DRV_MODULE_SYM qede |
| 36 | |
Sudarsana Kalluru | 133fac0 | 2015-10-26 11:02:34 +0200 | [diff] [blame] | 37 | struct qede_stats { |
| 38 | u64 no_buff_discards; |
| 39 | u64 rx_ucast_bytes; |
| 40 | u64 rx_mcast_bytes; |
| 41 | u64 rx_bcast_bytes; |
| 42 | u64 rx_ucast_pkts; |
| 43 | u64 rx_mcast_pkts; |
| 44 | u64 rx_bcast_pkts; |
| 45 | u64 mftag_filter_discards; |
| 46 | u64 mac_filter_discards; |
| 47 | u64 tx_ucast_bytes; |
| 48 | u64 tx_mcast_bytes; |
| 49 | u64 tx_bcast_bytes; |
| 50 | u64 tx_ucast_pkts; |
| 51 | u64 tx_mcast_pkts; |
| 52 | u64 tx_bcast_pkts; |
| 53 | u64 tx_err_drop_pkts; |
| 54 | u64 coalesced_pkts; |
| 55 | u64 coalesced_events; |
| 56 | u64 coalesced_aborts_num; |
| 57 | u64 non_coalesced_pkts; |
| 58 | u64 coalesced_bytes; |
| 59 | |
| 60 | /* port */ |
| 61 | u64 rx_64_byte_packets; |
| 62 | u64 rx_127_byte_packets; |
| 63 | u64 rx_255_byte_packets; |
| 64 | u64 rx_511_byte_packets; |
| 65 | u64 rx_1023_byte_packets; |
| 66 | u64 rx_1518_byte_packets; |
| 67 | u64 rx_1522_byte_packets; |
| 68 | u64 rx_2047_byte_packets; |
| 69 | u64 rx_4095_byte_packets; |
| 70 | u64 rx_9216_byte_packets; |
| 71 | u64 rx_16383_byte_packets; |
| 72 | u64 rx_crc_errors; |
| 73 | u64 rx_mac_crtl_frames; |
| 74 | u64 rx_pause_frames; |
| 75 | u64 rx_pfc_frames; |
| 76 | u64 rx_align_errors; |
| 77 | u64 rx_carrier_errors; |
| 78 | u64 rx_oversize_packets; |
| 79 | u64 rx_jabbers; |
| 80 | u64 rx_undersize_packets; |
| 81 | u64 rx_fragments; |
| 82 | u64 tx_64_byte_packets; |
| 83 | u64 tx_65_to_127_byte_packets; |
| 84 | u64 tx_128_to_255_byte_packets; |
| 85 | u64 tx_256_to_511_byte_packets; |
| 86 | u64 tx_512_to_1023_byte_packets; |
| 87 | u64 tx_1024_to_1518_byte_packets; |
| 88 | u64 tx_1519_to_2047_byte_packets; |
| 89 | u64 tx_2048_to_4095_byte_packets; |
| 90 | u64 tx_4096_to_9216_byte_packets; |
| 91 | u64 tx_9217_to_16383_byte_packets; |
| 92 | u64 tx_pause_frames; |
| 93 | u64 tx_pfc_frames; |
| 94 | u64 tx_lpi_entry_count; |
| 95 | u64 tx_total_collisions; |
| 96 | u64 brb_truncates; |
| 97 | u64 brb_discards; |
| 98 | u64 tx_mac_ctrl_frames; |
| 99 | }; |
| 100 | |
Sudarsana Reddy Kalluru | 7c1bfca | 2016-02-18 17:00:40 +0200 | [diff] [blame] | 101 | struct qede_vlan { |
| 102 | struct list_head list; |
| 103 | u16 vid; |
| 104 | bool configured; |
| 105 | }; |
| 106 | |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 107 | struct qede_dev { |
| 108 | struct qed_dev *cdev; |
| 109 | struct net_device *ndev; |
| 110 | struct pci_dev *pdev; |
| 111 | |
| 112 | u32 dp_module; |
| 113 | u8 dp_level; |
| 114 | |
| 115 | const struct qed_eth_ops *ops; |
| 116 | |
| 117 | struct qed_dev_eth_info dev_info; |
| 118 | #define QEDE_MAX_RSS_CNT(edev) ((edev)->dev_info.num_queues) |
| 119 | #define QEDE_MAX_TSS_CNT(edev) ((edev)->dev_info.num_queues * \ |
| 120 | (edev)->dev_info.num_tc) |
| 121 | |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 122 | struct qede_fastpath *fp_array; |
Sudarsana Kalluru | 8edf049 | 2015-11-30 12:25:01 +0200 | [diff] [blame] | 123 | u16 req_rss; |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 124 | u16 num_rss; |
| 125 | u8 num_tc; |
| 126 | #define QEDE_RSS_CNT(edev) ((edev)->num_rss) |
| 127 | #define QEDE_TSS_CNT(edev) ((edev)->num_rss * \ |
| 128 | (edev)->num_tc) |
| 129 | #define QEDE_TSS_IDX(edev, txqidx) ((txqidx) % (edev)->num_rss) |
| 130 | #define QEDE_TC_IDX(edev, txqidx) ((txqidx) / (edev)->num_rss) |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 131 | #define QEDE_TX_QUEUE(edev, txqidx) \ |
| 132 | (&(edev)->fp_array[QEDE_TSS_IDX((edev), (txqidx))].txqs[QEDE_TC_IDX( \ |
| 133 | (edev), (txqidx))]) |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 134 | |
| 135 | struct qed_int_info int_info; |
| 136 | unsigned char primary_mac[ETH_ALEN]; |
| 137 | |
| 138 | /* Smaller private varaiant of the RTNL lock */ |
| 139 | struct mutex qede_lock; |
| 140 | u32 state; /* Protected by qede_lock */ |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 141 | u16 rx_buf_size; |
| 142 | /* L2 header size + 2*VLANs (8 bytes) + LLC SNAP (8 bytes) */ |
| 143 | #define ETH_OVERHEAD (ETH_HLEN + 8 + 8) |
| 144 | /* Max supported alignment is 256 (8 shift) |
| 145 | * minimal alignment shift 6 is optimal for 57xxx HW performance |
| 146 | */ |
| 147 | #define QEDE_RX_ALIGN_SHIFT max(6, min(8, L1_CACHE_SHIFT)) |
| 148 | /* We assume skb_build() uses sizeof(struct skb_shared_info) bytes |
| 149 | * at the end of skb->data, to avoid wasting a full cache line. |
| 150 | * This reduces memory use (skb->truesize). |
| 151 | */ |
| 152 | #define QEDE_FW_RX_ALIGN_END \ |
| 153 | max_t(u64, 1UL << QEDE_RX_ALIGN_SHIFT, \ |
| 154 | SKB_DATA_ALIGN(sizeof(struct skb_shared_info))) |
| 155 | |
Sudarsana Kalluru | 133fac0 | 2015-10-26 11:02:34 +0200 | [diff] [blame] | 156 | struct qede_stats stats; |
Sudarsana Reddy Kalluru | 961acde | 2016-04-10 12:43:01 +0300 | [diff] [blame] | 157 | #define QEDE_RSS_INDIR_INITED BIT(0) |
| 158 | #define QEDE_RSS_KEY_INITED BIT(1) |
| 159 | #define QEDE_RSS_CAPS_INITED BIT(2) |
| 160 | u32 rss_params_inited; /* bit-field to track initialized rss params */ |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 161 | struct qed_update_vport_rss_params rss_params; |
| 162 | u16 q_num_rx_buffers; /* Must be a power of two */ |
| 163 | u16 q_num_tx_buffers; /* Must be a power of two */ |
Sudarsana Kalluru | 0d8e0aa | 2015-10-26 11:02:30 +0200 | [diff] [blame] | 164 | |
Manish Chopra | 55482ed | 2016-03-04 12:35:06 -0500 | [diff] [blame] | 165 | bool gro_disable; |
Sudarsana Reddy Kalluru | 7c1bfca | 2016-02-18 17:00:40 +0200 | [diff] [blame] | 166 | struct list_head vlan_list; |
| 167 | u16 configured_vlans; |
| 168 | u16 non_configured_vlans; |
| 169 | bool accept_any_vlan; |
Sudarsana Kalluru | 0d8e0aa | 2015-10-26 11:02:30 +0200 | [diff] [blame] | 170 | struct delayed_work sp_task; |
| 171 | unsigned long sp_flags; |
Manish Chopra | b18e170 | 2016-04-14 01:38:30 -0400 | [diff] [blame^] | 172 | u16 vxlan_dst_port; |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 173 | }; |
| 174 | |
| 175 | enum QEDE_STATE { |
| 176 | QEDE_STATE_CLOSED, |
| 177 | QEDE_STATE_OPEN, |
| 178 | }; |
| 179 | |
| 180 | #define HILO_U64(hi, lo) ((((u64)(hi)) << 32) + (lo)) |
| 181 | |
| 182 | #define MAX_NUM_TC 8 |
| 183 | #define MAX_NUM_PRI 8 |
| 184 | |
| 185 | /* The driver supports the new build_skb() API: |
| 186 | * RX ring buffer contains pointer to kmalloc() data only, |
| 187 | * skb are built only after the frame was DMA-ed. |
| 188 | */ |
| 189 | struct sw_rx_data { |
Yuval Mintz | fc48b7a | 2016-02-15 13:22:35 -0500 | [diff] [blame] | 190 | struct page *data; |
| 191 | dma_addr_t mapping; |
| 192 | unsigned int page_offset; |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 193 | }; |
| 194 | |
Manish Chopra | 55482ed | 2016-03-04 12:35:06 -0500 | [diff] [blame] | 195 | enum qede_agg_state { |
| 196 | QEDE_AGG_STATE_NONE = 0, |
| 197 | QEDE_AGG_STATE_START = 1, |
| 198 | QEDE_AGG_STATE_ERROR = 2 |
| 199 | }; |
| 200 | |
| 201 | struct qede_agg_info { |
| 202 | struct sw_rx_data replace_buf; |
| 203 | dma_addr_t replace_buf_mapping; |
| 204 | struct sw_rx_data start_buf; |
| 205 | dma_addr_t start_buf_mapping; |
| 206 | struct eth_fast_path_rx_tpa_start_cqe start_cqe; |
| 207 | enum qede_agg_state agg_state; |
| 208 | struct sk_buff *skb; |
| 209 | int frag_id; |
| 210 | u16 vlan_tag; |
| 211 | }; |
| 212 | |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 213 | struct qede_rx_queue { |
| 214 | __le16 *hw_cons_ptr; |
| 215 | struct sw_rx_data *sw_rx_ring; |
| 216 | u16 sw_rx_cons; |
| 217 | u16 sw_rx_prod; |
| 218 | struct qed_chain rx_bd_ring; |
| 219 | struct qed_chain rx_comp_ring; |
| 220 | void __iomem *hw_rxq_prod_addr; |
| 221 | |
Manish Chopra | 55482ed | 2016-03-04 12:35:06 -0500 | [diff] [blame] | 222 | /* GRO */ |
| 223 | struct qede_agg_info tpa_info[ETH_TPA_MAX_AGGS_NUM]; |
| 224 | |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 225 | int rx_buf_size; |
Yuval Mintz | fc48b7a | 2016-02-15 13:22:35 -0500 | [diff] [blame] | 226 | unsigned int rx_buf_seg_size; |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 227 | |
| 228 | u16 num_rx_buffers; |
| 229 | u16 rxq_id; |
| 230 | |
| 231 | u64 rx_hw_errors; |
| 232 | u64 rx_alloc_errors; |
| 233 | }; |
| 234 | |
| 235 | union db_prod { |
| 236 | struct eth_db_data data; |
| 237 | u32 raw; |
| 238 | }; |
| 239 | |
| 240 | struct sw_tx_bd { |
| 241 | struct sk_buff *skb; |
| 242 | u8 flags; |
| 243 | /* Set on the first BD descriptor when there is a split BD */ |
| 244 | #define QEDE_TSO_SPLIT_BD BIT(0) |
| 245 | }; |
| 246 | |
| 247 | struct qede_tx_queue { |
| 248 | int index; /* Queue index */ |
| 249 | __le16 *hw_cons_ptr; |
| 250 | struct sw_tx_bd *sw_tx_ring; |
| 251 | u16 sw_tx_cons; |
| 252 | u16 sw_tx_prod; |
| 253 | struct qed_chain tx_pbl; |
| 254 | void __iomem *doorbell_addr; |
| 255 | union db_prod tx_db; |
| 256 | |
| 257 | u16 num_tx_buffers; |
| 258 | }; |
| 259 | |
| 260 | #define BD_UNMAP_ADDR(bd) HILO_U64(le32_to_cpu((bd)->addr.hi), \ |
| 261 | le32_to_cpu((bd)->addr.lo)) |
| 262 | #define BD_SET_UNMAP_ADDR_LEN(bd, maddr, len) \ |
| 263 | do { \ |
| 264 | (bd)->addr.hi = cpu_to_le32(upper_32_bits(maddr)); \ |
| 265 | (bd)->addr.lo = cpu_to_le32(lower_32_bits(maddr)); \ |
| 266 | (bd)->nbytes = cpu_to_le16(len); \ |
| 267 | } while (0) |
| 268 | #define BD_UNMAP_LEN(bd) (le16_to_cpu((bd)->nbytes)) |
| 269 | |
| 270 | struct qede_fastpath { |
| 271 | struct qede_dev *edev; |
| 272 | u8 rss_id; |
| 273 | struct napi_struct napi; |
| 274 | struct qed_sb_info *sb_info; |
| 275 | struct qede_rx_queue *rxq; |
| 276 | struct qede_tx_queue *txqs; |
| 277 | |
| 278 | #define VEC_NAME_SIZE (sizeof(((struct net_device *)0)->name) + 8) |
| 279 | char name[VEC_NAME_SIZE]; |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 280 | }; |
| 281 | |
| 282 | /* Debug print definitions */ |
| 283 | #define DP_NAME(edev) ((edev)->ndev->name) |
| 284 | |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 285 | #define XMIT_PLAIN 0 |
| 286 | #define XMIT_L4_CSUM BIT(0) |
| 287 | #define XMIT_LSO BIT(1) |
| 288 | #define XMIT_ENC BIT(2) |
| 289 | |
| 290 | #define QEDE_CSUM_ERROR BIT(0) |
| 291 | #define QEDE_CSUM_UNNECESSARY BIT(1) |
Sudarsana Kalluru | 0d8e0aa | 2015-10-26 11:02:30 +0200 | [diff] [blame] | 292 | |
Manish Chopra | b18e170 | 2016-04-14 01:38:30 -0400 | [diff] [blame^] | 293 | #define QEDE_SP_RX_MODE 1 |
| 294 | #define QEDE_SP_VXLAN_PORT_CONFIG 2 |
Sudarsana Kalluru | 0d8e0aa | 2015-10-26 11:02:30 +0200 | [diff] [blame] | 295 | |
| 296 | union qede_reload_args { |
| 297 | u16 mtu; |
| 298 | }; |
| 299 | |
Sudarsana Kalluru | 133fac0 | 2015-10-26 11:02:34 +0200 | [diff] [blame] | 300 | void qede_config_debug(uint debug, u32 *p_dp_module, u8 *p_dp_level); |
| 301 | void qede_set_ethtool_ops(struct net_device *netdev); |
| 302 | void qede_reload(struct qede_dev *edev, |
| 303 | void (*func)(struct qede_dev *edev, |
| 304 | union qede_reload_args *args), |
| 305 | union qede_reload_args *args); |
| 306 | int qede_change_mtu(struct net_device *dev, int new_mtu); |
| 307 | void qede_fill_by_demand_stats(struct qede_dev *edev); |
| 308 | |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 309 | #define RX_RING_SIZE_POW 13 |
Sudarsana Kalluru | 01ef7e0 | 2015-11-30 12:25:02 +0200 | [diff] [blame] | 310 | #define RX_RING_SIZE ((u16)BIT(RX_RING_SIZE_POW)) |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 311 | #define NUM_RX_BDS_MAX (RX_RING_SIZE - 1) |
| 312 | #define NUM_RX_BDS_MIN 128 |
| 313 | #define NUM_RX_BDS_DEF NUM_RX_BDS_MAX |
| 314 | |
| 315 | #define TX_RING_SIZE_POW 13 |
Sudarsana Kalluru | 01ef7e0 | 2015-11-30 12:25:02 +0200 | [diff] [blame] | 316 | #define TX_RING_SIZE ((u16)BIT(TX_RING_SIZE_POW)) |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 317 | #define NUM_TX_BDS_MAX (TX_RING_SIZE - 1) |
| 318 | #define NUM_TX_BDS_MIN 128 |
| 319 | #define NUM_TX_BDS_DEF NUM_TX_BDS_MAX |
| 320 | |
Yuval Mintz | fc48b7a | 2016-02-15 13:22:35 -0500 | [diff] [blame] | 321 | #define QEDE_RX_HDR_SIZE 256 |
Yuval Mintz | 2950219 | 2015-10-26 11:02:29 +0200 | [diff] [blame] | 322 | #define for_each_rss(i) for (i = 0; i < edev->num_rss; i++) |
| 323 | |
Yuval Mintz | e712d52 | 2015-10-26 11:02:27 +0200 | [diff] [blame] | 324 | #endif /* _QEDE_H_ */ |