Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Linux driver for VMware's vmxnet3 ethernet NIC. |
| 3 | * |
| 4 | * Copyright (C) 2008-2009, VMware, Inc. All Rights Reserved. |
| 5 | * |
| 6 | * This program is free software; you can redistribute it and/or modify it |
| 7 | * under the terms of the GNU General Public License as published by the |
| 8 | * Free Software Foundation; version 2 of the License and no later version. |
| 9 | * |
| 10 | * This program is distributed in the hope that it will be useful, but |
| 11 | * WITHOUT ANY WARRANTY; without even the implied warranty of |
| 12 | * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or |
| 13 | * NON INFRINGEMENT. See the GNU General Public License for more |
| 14 | * details. |
| 15 | * |
| 16 | * You should have received a copy of the GNU General Public License |
| 17 | * along with this program; if not, write to the Free Software |
| 18 | * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 19 | * |
| 20 | * The full GNU General Public License is included in this distribution in |
| 21 | * the file called "COPYING". |
| 22 | * |
| 23 | * Maintained by: Shreyas Bhatewara <pv-drivers@vmware.com> |
| 24 | * |
| 25 | */ |
| 26 | |
| 27 | #ifndef _VMXNET3_INT_H |
| 28 | #define _VMXNET3_INT_H |
| 29 | |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 30 | #include <linux/ethtool.h> |
| 31 | #include <linux/delay.h> |
| 32 | #include <linux/netdevice.h> |
| 33 | #include <linux/pci.h> |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 34 | #include <linux/compiler.h> |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 35 | #include <linux/slab.h> |
| 36 | #include <linux/spinlock.h> |
| 37 | #include <linux/ioport.h> |
| 38 | #include <linux/highmem.h> |
| 39 | #include <linux/init.h> |
| 40 | #include <linux/timer.h> |
| 41 | #include <linux/skbuff.h> |
| 42 | #include <linux/interrupt.h> |
| 43 | #include <linux/workqueue.h> |
| 44 | #include <linux/uaccess.h> |
| 45 | #include <asm/dma.h> |
| 46 | #include <asm/page.h> |
| 47 | |
| 48 | #include <linux/tcp.h> |
| 49 | #include <linux/udp.h> |
| 50 | #include <linux/ip.h> |
| 51 | #include <linux/ipv6.h> |
| 52 | #include <linux/in.h> |
| 53 | #include <linux/etherdevice.h> |
| 54 | #include <asm/checksum.h> |
| 55 | #include <linux/if_vlan.h> |
| 56 | #include <linux/if_arp.h> |
| 57 | #include <linux/inetdevice.h> |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 58 | |
| 59 | #include "vmxnet3_defs.h" |
| 60 | |
| 61 | #ifdef DEBUG |
| 62 | # define VMXNET3_DRIVER_VERSION_REPORT VMXNET3_DRIVER_VERSION_STRING"-NAPI(debug)" |
| 63 | #else |
| 64 | # define VMXNET3_DRIVER_VERSION_REPORT VMXNET3_DRIVER_VERSION_STRING"-NAPI" |
| 65 | #endif |
| 66 | |
| 67 | |
| 68 | /* |
| 69 | * Version numbers |
| 70 | */ |
Bhavesh Davda | 1f4b161 | 2010-07-24 14:43:29 +0000 | [diff] [blame] | 71 | #define VMXNET3_DRIVER_VERSION_STRING "1.0.14.0-k" |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 72 | |
| 73 | /* a 32-bit int, each byte encode a verion number in VMXNET3_DRIVER_VERSION */ |
Bhavesh Davda | 1f4b161 | 2010-07-24 14:43:29 +0000 | [diff] [blame] | 74 | #define VMXNET3_DRIVER_VERSION_NUM 0x01000E00 |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 75 | |
| 76 | |
| 77 | /* |
| 78 | * Capabilities |
| 79 | */ |
| 80 | |
| 81 | enum { |
| 82 | VMNET_CAP_SG = 0x0001, /* Can do scatter-gather transmits. */ |
| 83 | VMNET_CAP_IP4_CSUM = 0x0002, /* Can checksum only TCP/UDP over |
| 84 | * IPv4 */ |
| 85 | VMNET_CAP_HW_CSUM = 0x0004, /* Can checksum all packets. */ |
| 86 | VMNET_CAP_HIGH_DMA = 0x0008, /* Can DMA to high memory. */ |
| 87 | VMNET_CAP_TOE = 0x0010, /* Supports TCP/IP offload. */ |
| 88 | VMNET_CAP_TSO = 0x0020, /* Supports TCP Segmentation |
| 89 | * offload */ |
| 90 | VMNET_CAP_SW_TSO = 0x0040, /* Supports SW TCP Segmentation */ |
| 91 | VMNET_CAP_VMXNET_APROM = 0x0080, /* Vmxnet APROM support */ |
| 92 | VMNET_CAP_HW_TX_VLAN = 0x0100, /* Can we do VLAN tagging in HW */ |
| 93 | VMNET_CAP_HW_RX_VLAN = 0x0200, /* Can we do VLAN untagging in HW */ |
| 94 | VMNET_CAP_SW_VLAN = 0x0400, /* VLAN tagging/untagging in SW */ |
| 95 | VMNET_CAP_WAKE_PCKT_RCV = 0x0800, /* Can wake on network packet recv? */ |
| 96 | VMNET_CAP_ENABLE_INT_INLINE = 0x1000, /* Enable Interrupt Inline */ |
| 97 | VMNET_CAP_ENABLE_HEADER_COPY = 0x2000, /* copy header for vmkernel */ |
| 98 | VMNET_CAP_TX_CHAIN = 0x4000, /* Guest can use multiple tx entries |
| 99 | * for a pkt */ |
| 100 | VMNET_CAP_RX_CHAIN = 0x8000, /* pkt can span multiple rx entries */ |
| 101 | VMNET_CAP_LPD = 0x10000, /* large pkt delivery */ |
| 102 | VMNET_CAP_BPF = 0x20000, /* BPF Support in VMXNET Virtual HW*/ |
| 103 | VMNET_CAP_SG_SPAN_PAGES = 0x40000, /* Scatter-gather can span multiple*/ |
| 104 | /* pages transmits */ |
| 105 | VMNET_CAP_IP6_CSUM = 0x80000, /* Can do IPv6 csum offload. */ |
| 106 | VMNET_CAP_TSO6 = 0x100000, /* TSO seg. offload for IPv6 pkts. */ |
| 107 | VMNET_CAP_TSO256k = 0x200000, /* Can do TSO seg offload for */ |
| 108 | /* pkts up to 256kB. */ |
| 109 | VMNET_CAP_UPT = 0x400000 /* Support UPT */ |
| 110 | }; |
| 111 | |
| 112 | /* |
| 113 | * PCI vendor and device IDs. |
| 114 | */ |
| 115 | #define PCI_VENDOR_ID_VMWARE 0x15AD |
| 116 | #define PCI_DEVICE_ID_VMWARE_VMXNET3 0x07B0 |
| 117 | #define MAX_ETHERNET_CARDS 10 |
| 118 | #define MAX_PCI_PASSTHRU_DEVICE 6 |
| 119 | |
| 120 | struct vmxnet3_cmd_ring { |
| 121 | union Vmxnet3_GenericDesc *base; |
| 122 | u32 size; |
| 123 | u32 next2fill; |
| 124 | u32 next2comp; |
| 125 | u8 gen; |
| 126 | dma_addr_t basePA; |
| 127 | }; |
| 128 | |
| 129 | static inline void |
| 130 | vmxnet3_cmd_ring_adv_next2fill(struct vmxnet3_cmd_ring *ring) |
| 131 | { |
| 132 | ring->next2fill++; |
| 133 | if (unlikely(ring->next2fill == ring->size)) { |
| 134 | ring->next2fill = 0; |
| 135 | VMXNET3_FLIP_RING_GEN(ring->gen); |
| 136 | } |
| 137 | } |
| 138 | |
| 139 | static inline void |
| 140 | vmxnet3_cmd_ring_adv_next2comp(struct vmxnet3_cmd_ring *ring) |
| 141 | { |
| 142 | VMXNET3_INC_RING_IDX_ONLY(ring->next2comp, ring->size); |
| 143 | } |
| 144 | |
| 145 | static inline int |
| 146 | vmxnet3_cmd_ring_desc_avail(struct vmxnet3_cmd_ring *ring) |
| 147 | { |
| 148 | return (ring->next2comp > ring->next2fill ? 0 : ring->size) + |
| 149 | ring->next2comp - ring->next2fill - 1; |
| 150 | } |
| 151 | |
| 152 | struct vmxnet3_comp_ring { |
| 153 | union Vmxnet3_GenericDesc *base; |
| 154 | u32 size; |
| 155 | u32 next2proc; |
| 156 | u8 gen; |
| 157 | u8 intr_idx; |
| 158 | dma_addr_t basePA; |
| 159 | }; |
| 160 | |
| 161 | static inline void |
| 162 | vmxnet3_comp_ring_adv_next2proc(struct vmxnet3_comp_ring *ring) |
| 163 | { |
| 164 | ring->next2proc++; |
| 165 | if (unlikely(ring->next2proc == ring->size)) { |
| 166 | ring->next2proc = 0; |
| 167 | VMXNET3_FLIP_RING_GEN(ring->gen); |
| 168 | } |
| 169 | } |
| 170 | |
| 171 | struct vmxnet3_tx_data_ring { |
| 172 | struct Vmxnet3_TxDataDesc *base; |
| 173 | u32 size; |
| 174 | dma_addr_t basePA; |
| 175 | }; |
| 176 | |
| 177 | enum vmxnet3_buf_map_type { |
| 178 | VMXNET3_MAP_INVALID = 0, |
| 179 | VMXNET3_MAP_NONE, |
| 180 | VMXNET3_MAP_SINGLE, |
| 181 | VMXNET3_MAP_PAGE, |
| 182 | }; |
| 183 | |
| 184 | struct vmxnet3_tx_buf_info { |
| 185 | u32 map_type; |
| 186 | u16 len; |
| 187 | u16 sop_idx; |
| 188 | dma_addr_t dma_addr; |
| 189 | struct sk_buff *skb; |
| 190 | }; |
| 191 | |
| 192 | struct vmxnet3_tq_driver_stats { |
| 193 | u64 drop_total; /* # of pkts dropped by the driver, the |
| 194 | * counters below track droppings due to |
| 195 | * different reasons |
| 196 | */ |
| 197 | u64 drop_too_many_frags; |
| 198 | u64 drop_oversized_hdr; |
| 199 | u64 drop_hdr_inspect_err; |
| 200 | u64 drop_tso; |
| 201 | |
| 202 | u64 tx_ring_full; |
| 203 | u64 linearized; /* # of pkts linearized */ |
| 204 | u64 copy_skb_header; /* # of times we have to copy skb header */ |
| 205 | u64 oversized_hdr; |
| 206 | }; |
| 207 | |
| 208 | struct vmxnet3_tx_ctx { |
| 209 | bool ipv4; |
| 210 | u16 mss; |
| 211 | u32 eth_ip_hdr_size; /* only valid for pkts requesting tso or csum |
| 212 | * offloading |
| 213 | */ |
| 214 | u32 l4_hdr_size; /* only valid if mss != 0 */ |
| 215 | u32 copy_size; /* # of bytes copied into the data ring */ |
| 216 | union Vmxnet3_GenericDesc *sop_txd; |
| 217 | union Vmxnet3_GenericDesc *eop_txd; |
| 218 | }; |
| 219 | |
| 220 | struct vmxnet3_tx_queue { |
| 221 | spinlock_t tx_lock; |
| 222 | struct vmxnet3_cmd_ring tx_ring; |
| 223 | struct vmxnet3_tx_buf_info *buf_info; |
| 224 | struct vmxnet3_tx_data_ring data_ring; |
| 225 | struct vmxnet3_comp_ring comp_ring; |
| 226 | struct Vmxnet3_TxQueueCtrl *shared; |
| 227 | struct vmxnet3_tq_driver_stats stats; |
| 228 | bool stopped; |
| 229 | int num_stop; /* # of times the queue is |
| 230 | * stopped */ |
| 231 | } __attribute__((__aligned__(SMP_CACHE_BYTES))); |
| 232 | |
| 233 | enum vmxnet3_rx_buf_type { |
| 234 | VMXNET3_RX_BUF_NONE = 0, |
| 235 | VMXNET3_RX_BUF_SKB = 1, |
| 236 | VMXNET3_RX_BUF_PAGE = 2 |
| 237 | }; |
| 238 | |
| 239 | struct vmxnet3_rx_buf_info { |
| 240 | enum vmxnet3_rx_buf_type buf_type; |
| 241 | u16 len; |
| 242 | union { |
| 243 | struct sk_buff *skb; |
| 244 | struct page *page; |
| 245 | }; |
| 246 | dma_addr_t dma_addr; |
| 247 | }; |
| 248 | |
| 249 | struct vmxnet3_rx_ctx { |
| 250 | struct sk_buff *skb; |
| 251 | u32 sop_idx; |
| 252 | }; |
| 253 | |
| 254 | struct vmxnet3_rq_driver_stats { |
| 255 | u64 drop_total; |
| 256 | u64 drop_err; |
| 257 | u64 drop_fcs; |
| 258 | u64 rx_buf_alloc_failure; |
| 259 | }; |
| 260 | |
| 261 | struct vmxnet3_rx_queue { |
| 262 | struct vmxnet3_cmd_ring rx_ring[2]; |
| 263 | struct vmxnet3_comp_ring comp_ring; |
| 264 | struct vmxnet3_rx_ctx rx_ctx; |
| 265 | u32 qid; /* rqID in RCD for buffer from 1st ring */ |
| 266 | u32 qid2; /* rqID in RCD for buffer from 2nd ring */ |
| 267 | u32 uncommitted[2]; /* # of buffers allocated since last RXPROD |
| 268 | * update */ |
| 269 | struct vmxnet3_rx_buf_info *buf_info[2]; |
| 270 | struct Vmxnet3_RxQueueCtrl *shared; |
| 271 | struct vmxnet3_rq_driver_stats stats; |
| 272 | } __attribute__((__aligned__(SMP_CACHE_BYTES))); |
| 273 | |
| 274 | #define VMXNET3_LINUX_MAX_MSIX_VECT 1 |
| 275 | |
| 276 | struct vmxnet3_intr { |
| 277 | enum vmxnet3_intr_mask_mode mask_mode; |
| 278 | enum vmxnet3_intr_type type; /* MSI-X, MSI, or INTx? */ |
| 279 | u8 num_intrs; /* # of intr vectors */ |
| 280 | u8 event_intr_idx; /* idx of the intr vector for event */ |
| 281 | u8 mod_levels[VMXNET3_LINUX_MAX_MSIX_VECT]; /* moderation level */ |
| 282 | #ifdef CONFIG_PCI_MSI |
| 283 | struct msix_entry msix_entries[VMXNET3_LINUX_MAX_MSIX_VECT]; |
| 284 | #endif |
| 285 | }; |
| 286 | |
| 287 | #define VMXNET3_STATE_BIT_RESETTING 0 |
| 288 | #define VMXNET3_STATE_BIT_QUIESCED 1 |
| 289 | struct vmxnet3_adapter { |
| 290 | struct vmxnet3_tx_queue tx_queue; |
| 291 | struct vmxnet3_rx_queue rx_queue; |
| 292 | struct napi_struct napi; |
| 293 | struct vlan_group *vlan_grp; |
| 294 | |
| 295 | struct vmxnet3_intr intr; |
| 296 | |
| 297 | struct Vmxnet3_DriverShared *shared; |
| 298 | struct Vmxnet3_PMConf *pm_conf; |
| 299 | struct Vmxnet3_TxQueueDesc *tqd_start; /* first tx queue desc */ |
| 300 | struct Vmxnet3_RxQueueDesc *rqd_start; /* first rx queue desc */ |
| 301 | struct net_device *netdev; |
| 302 | struct pci_dev *pdev; |
| 303 | |
| 304 | u8 *hw_addr0; /* for BAR 0 */ |
| 305 | u8 *hw_addr1; /* for BAR 1 */ |
| 306 | |
| 307 | /* feature control */ |
| 308 | bool rxcsum; |
| 309 | bool lro; |
| 310 | bool jumbo_frame; |
| 311 | |
| 312 | /* rx buffer related */ |
| 313 | unsigned skb_buf_size; |
| 314 | int rx_buf_per_pkt; /* only apply to the 1st ring */ |
| 315 | dma_addr_t shared_pa; |
| 316 | dma_addr_t queue_desc_pa; |
| 317 | |
| 318 | /* Wake-on-LAN */ |
| 319 | u32 wol; |
| 320 | |
| 321 | /* Link speed */ |
| 322 | u32 link_speed; /* in mbps */ |
| 323 | |
| 324 | u64 tx_timeout_count; |
| 325 | struct work_struct work; |
| 326 | |
| 327 | unsigned long state; /* VMXNET3_STATE_BIT_xxx */ |
| 328 | |
| 329 | int dev_number; |
| 330 | }; |
| 331 | |
| 332 | #define VMXNET3_WRITE_BAR0_REG(adapter, reg, val) \ |
Shreyas Bhatewara | 115924b | 2009-11-16 13:41:33 +0000 | [diff] [blame] | 333 | writel(cpu_to_le32(val), (adapter)->hw_addr0 + (reg)) |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 334 | #define VMXNET3_READ_BAR0_REG(adapter, reg) \ |
Shreyas Bhatewara | 115924b | 2009-11-16 13:41:33 +0000 | [diff] [blame] | 335 | le32_to_cpu(readl((adapter)->hw_addr0 + (reg))) |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 336 | |
| 337 | #define VMXNET3_WRITE_BAR1_REG(adapter, reg, val) \ |
Shreyas Bhatewara | 115924b | 2009-11-16 13:41:33 +0000 | [diff] [blame] | 338 | writel(cpu_to_le32(val), (adapter)->hw_addr1 + (reg)) |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 339 | #define VMXNET3_READ_BAR1_REG(adapter, reg) \ |
Shreyas Bhatewara | 115924b | 2009-11-16 13:41:33 +0000 | [diff] [blame] | 340 | le32_to_cpu(readl((adapter)->hw_addr1 + (reg))) |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 341 | |
| 342 | #define VMXNET3_WAKE_QUEUE_THRESHOLD(tq) (5) |
| 343 | #define VMXNET3_RX_ALLOC_THRESHOLD(rq, ring_idx, adapter) \ |
| 344 | ((rq)->rx_ring[ring_idx].size >> 3) |
| 345 | |
| 346 | #define VMXNET3_GET_ADDR_LO(dma) ((u32)(dma)) |
| 347 | #define VMXNET3_GET_ADDR_HI(dma) ((u32)(((u64)(dma)) >> 32)) |
| 348 | |
| 349 | /* must be a multiple of VMXNET3_RING_SIZE_ALIGN */ |
| 350 | #define VMXNET3_DEF_TX_RING_SIZE 512 |
| 351 | #define VMXNET3_DEF_RX_RING_SIZE 256 |
| 352 | |
| 353 | #define VMXNET3_MAX_ETH_HDR_SIZE 22 |
| 354 | #define VMXNET3_MAX_SKB_BUF_SIZE (3*1024) |
| 355 | |
Shreyas Bhatewara | 115924b | 2009-11-16 13:41:33 +0000 | [diff] [blame] | 356 | void set_flag_le16(__le16 *data, u16 flag); |
| 357 | void set_flag_le64(__le64 *data, u64 flag); |
| 358 | void reset_flag_le64(__le64 *data, u64 flag); |
| 359 | |
Shreyas Bhatewara | d1a890fa | 2009-10-13 00:15:51 -0700 | [diff] [blame] | 360 | int |
| 361 | vmxnet3_quiesce_dev(struct vmxnet3_adapter *adapter); |
| 362 | |
| 363 | int |
| 364 | vmxnet3_activate_dev(struct vmxnet3_adapter *adapter); |
| 365 | |
| 366 | void |
| 367 | vmxnet3_force_close(struct vmxnet3_adapter *adapter); |
| 368 | |
| 369 | void |
| 370 | vmxnet3_reset_dev(struct vmxnet3_adapter *adapter); |
| 371 | |
| 372 | void |
| 373 | vmxnet3_tq_destroy(struct vmxnet3_tx_queue *tq, |
| 374 | struct vmxnet3_adapter *adapter); |
| 375 | |
| 376 | void |
| 377 | vmxnet3_rq_destroy(struct vmxnet3_rx_queue *rq, |
| 378 | struct vmxnet3_adapter *adapter); |
| 379 | |
| 380 | int |
| 381 | vmxnet3_create_queues(struct vmxnet3_adapter *adapter, |
| 382 | u32 tx_ring_size, u32 rx_ring_size, u32 rx_ring2_size); |
| 383 | |
| 384 | extern void vmxnet3_set_ethtool_ops(struct net_device *netdev); |
| 385 | extern struct net_device_stats *vmxnet3_get_stats(struct net_device *netdev); |
| 386 | |
| 387 | extern char vmxnet3_driver_name[]; |
| 388 | #endif |