Jeff Kirsher | ae06c70 | 2018-03-22 10:08:48 -0700 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
Jacob Keller | 8664109 | 2016-04-07 08:21:21 -0700 | [diff] [blame] | 2 | /* Intel(R) Ethernet Switch Host Interface Driver |
Jacob Keller | 4abf01b | 2017-07-10 13:23:09 -0700 | [diff] [blame] | 3 | * Copyright(c) 2013 - 2017 Intel Corporation. |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 4 | * |
| 5 | * This program is free software; you can redistribute it and/or modify it |
| 6 | * under the terms and conditions of the GNU General Public License, |
| 7 | * version 2, as published by the Free Software Foundation. |
| 8 | * |
| 9 | * This program is distributed in the hope it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| 12 | * more details. |
| 13 | * |
| 14 | * The full GNU General Public License is included in this distribution in |
| 15 | * the file called "COPYING". |
| 16 | * |
| 17 | * Contact Information: |
| 18 | * e1000-devel Mailing List <e1000-devel@lists.sourceforge.net> |
| 19 | * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 |
| 20 | */ |
| 21 | |
| 22 | #include "fm10k.h" |
| 23 | #include "fm10k_vf.h" |
| 24 | #include "fm10k_pf.h" |
| 25 | |
| 26 | static s32 fm10k_iov_msg_error(struct fm10k_hw *hw, u32 **results, |
| 27 | struct fm10k_mbx_info *mbx) |
| 28 | { |
| 29 | struct fm10k_vf_info *vf_info = (struct fm10k_vf_info *)mbx; |
| 30 | struct fm10k_intfc *interface = hw->back; |
| 31 | struct pci_dev *pdev = interface->pdev; |
| 32 | |
| 33 | dev_err(&pdev->dev, "Unknown message ID %u on VF %d\n", |
| 34 | **results & FM10K_TLV_ID_MASK, vf_info->vf_idx); |
| 35 | |
| 36 | return fm10k_tlv_msg_error(hw, results, mbx); |
| 37 | } |
| 38 | |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 39 | /** |
| 40 | * fm10k_iov_msg_queue_mac_vlan - Message handler for MAC/VLAN request from VF |
| 41 | * @hw: Pointer to hardware structure |
| 42 | * @results: Pointer array to message, results[0] is pointer to message |
| 43 | * @mbx: Pointer to mailbox information structure |
| 44 | * |
| 45 | * This function is a custom handler for MAC/VLAN requests from the VF. The |
| 46 | * assumption is that it is acceptable to directly hand off the message from |
| 47 | * the VF to the PF's switch manager. However, we use a MAC/VLAN message |
| 48 | * queue to avoid overloading the mailbox when a large number of requests |
| 49 | * come in. |
| 50 | **/ |
| 51 | static s32 fm10k_iov_msg_queue_mac_vlan(struct fm10k_hw *hw, u32 **results, |
| 52 | struct fm10k_mbx_info *mbx) |
| 53 | { |
| 54 | struct fm10k_vf_info *vf_info = (struct fm10k_vf_info *)mbx; |
| 55 | struct fm10k_intfc *interface = hw->back; |
| 56 | u8 mac[ETH_ALEN]; |
| 57 | u32 *result; |
| 58 | int err = 0; |
| 59 | bool set; |
| 60 | u16 vlan; |
| 61 | u32 vid; |
| 62 | |
| 63 | /* we shouldn't be updating rules on a disabled interface */ |
| 64 | if (!FM10K_VF_FLAG_ENABLED(vf_info)) |
| 65 | err = FM10K_ERR_PARAM; |
| 66 | |
| 67 | if (!err && !!results[FM10K_MAC_VLAN_MSG_VLAN]) { |
| 68 | result = results[FM10K_MAC_VLAN_MSG_VLAN]; |
| 69 | |
| 70 | /* record VLAN id requested */ |
| 71 | err = fm10k_tlv_attr_get_u32(result, &vid); |
| 72 | if (err) |
| 73 | return err; |
| 74 | |
| 75 | set = !(vid & FM10K_VLAN_CLEAR); |
| 76 | vid &= ~FM10K_VLAN_CLEAR; |
| 77 | |
| 78 | /* if the length field has been set, this is a multi-bit |
| 79 | * update request. For multi-bit requests, simply disallow |
| 80 | * them when the pf_vid has been set. In this case, the PF |
| 81 | * should have already cleared the VLAN_TABLE, and if we |
| 82 | * allowed them, it could allow a rogue VF to receive traffic |
| 83 | * on a VLAN it was not assigned. In the single-bit case, we |
| 84 | * need to modify requests for VLAN 0 to use the default PF or |
| 85 | * SW vid when assigned. |
| 86 | */ |
| 87 | |
| 88 | if (vid >> 16) { |
| 89 | /* prevent multi-bit requests when PF has |
| 90 | * administratively set the VLAN for this VF |
| 91 | */ |
| 92 | if (vf_info->pf_vid) |
| 93 | return FM10K_ERR_PARAM; |
| 94 | } else { |
| 95 | err = fm10k_iov_select_vid(vf_info, (u16)vid); |
| 96 | if (err < 0) |
| 97 | return err; |
| 98 | |
| 99 | vid = err; |
| 100 | } |
| 101 | |
| 102 | /* update VSI info for VF in regards to VLAN table */ |
| 103 | err = hw->mac.ops.update_vlan(hw, vid, vf_info->vsi, set); |
| 104 | } |
| 105 | |
| 106 | if (!err && !!results[FM10K_MAC_VLAN_MSG_MAC]) { |
| 107 | result = results[FM10K_MAC_VLAN_MSG_MAC]; |
| 108 | |
| 109 | /* record unicast MAC address requested */ |
| 110 | err = fm10k_tlv_attr_get_mac_vlan(result, mac, &vlan); |
| 111 | if (err) |
| 112 | return err; |
| 113 | |
| 114 | /* block attempts to set MAC for a locked device */ |
| 115 | if (is_valid_ether_addr(vf_info->mac) && |
| 116 | !ether_addr_equal(mac, vf_info->mac)) |
| 117 | return FM10K_ERR_PARAM; |
| 118 | |
| 119 | set = !(vlan & FM10K_VLAN_CLEAR); |
| 120 | vlan &= ~FM10K_VLAN_CLEAR; |
| 121 | |
| 122 | err = fm10k_iov_select_vid(vf_info, vlan); |
| 123 | if (err < 0) |
| 124 | return err; |
| 125 | |
| 126 | vlan = (u16)err; |
| 127 | |
| 128 | /* Add this request to the MAC/VLAN queue */ |
| 129 | err = fm10k_queue_mac_request(interface, vf_info->glort, |
| 130 | mac, vlan, set); |
| 131 | } |
| 132 | |
| 133 | if (!err && !!results[FM10K_MAC_VLAN_MSG_MULTICAST]) { |
| 134 | result = results[FM10K_MAC_VLAN_MSG_MULTICAST]; |
| 135 | |
| 136 | /* record multicast MAC address requested */ |
| 137 | err = fm10k_tlv_attr_get_mac_vlan(result, mac, &vlan); |
| 138 | if (err) |
| 139 | return err; |
| 140 | |
| 141 | /* verify that the VF is allowed to request multicast */ |
| 142 | if (!(vf_info->vf_flags & FM10K_VF_FLAG_MULTI_ENABLED)) |
| 143 | return FM10K_ERR_PARAM; |
| 144 | |
| 145 | set = !(vlan & FM10K_VLAN_CLEAR); |
| 146 | vlan &= ~FM10K_VLAN_CLEAR; |
| 147 | |
| 148 | err = fm10k_iov_select_vid(vf_info, vlan); |
| 149 | if (err < 0) |
| 150 | return err; |
| 151 | |
| 152 | vlan = (u16)err; |
| 153 | |
| 154 | /* Add this request to the MAC/VLAN queue */ |
| 155 | err = fm10k_queue_mac_request(interface, vf_info->glort, |
| 156 | mac, vlan, set); |
| 157 | } |
| 158 | |
| 159 | return err; |
| 160 | } |
| 161 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 162 | static const struct fm10k_msg_data iov_mbx_data[] = { |
| 163 | FM10K_TLV_MSG_TEST_HANDLER(fm10k_tlv_msg_test), |
| 164 | FM10K_VF_MSG_MSIX_HANDLER(fm10k_iov_msg_msix_pf), |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 165 | FM10K_VF_MSG_MAC_VLAN_HANDLER(fm10k_iov_msg_queue_mac_vlan), |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 166 | FM10K_VF_MSG_LPORT_STATE_HANDLER(fm10k_iov_msg_lport_state_pf), |
| 167 | FM10K_TLV_MSG_ERROR_HANDLER(fm10k_iov_msg_error), |
| 168 | }; |
| 169 | |
| 170 | s32 fm10k_iov_event(struct fm10k_intfc *interface) |
| 171 | { |
| 172 | struct fm10k_hw *hw = &interface->hw; |
| 173 | struct fm10k_iov_data *iov_data; |
Jeff Kirsher | 9de15bd | 2015-04-10 17:20:17 -0700 | [diff] [blame] | 174 | s64 vflre; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 175 | int i; |
| 176 | |
Jacob Keller | d8ec92f | 2016-02-10 14:45:51 -0800 | [diff] [blame] | 177 | /* if there is no iov_data then there is no mailbox to process */ |
Jacob Keller | ce4dad2 | 2016-06-17 16:21:11 -0700 | [diff] [blame] | 178 | if (!READ_ONCE(interface->iov_data)) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 179 | return 0; |
| 180 | |
| 181 | rcu_read_lock(); |
| 182 | |
| 183 | iov_data = interface->iov_data; |
| 184 | |
| 185 | /* check again now that we are in the RCU block */ |
| 186 | if (!iov_data) |
| 187 | goto read_unlock; |
| 188 | |
| 189 | if (!(fm10k_read_reg(hw, FM10K_EICR) & FM10K_EICR_VFLR)) |
Jeff Kirsher | 9de15bd | 2015-04-10 17:20:17 -0700 | [diff] [blame] | 190 | goto read_unlock; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 191 | |
| 192 | /* read VFLRE to determine if any VFs have been reset */ |
Jacob Keller | d876c15 | 2017-07-10 13:23:10 -0700 | [diff] [blame] | 193 | vflre = fm10k_read_reg(hw, FM10K_PFVFLRE(1)); |
| 194 | vflre <<= 32; |
| 195 | vflre |= fm10k_read_reg(hw, FM10K_PFVFLRE(0)); |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 196 | |
Jacob Keller | d876c15 | 2017-07-10 13:23:10 -0700 | [diff] [blame] | 197 | i = iov_data->num_vfs; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 198 | |
Jacob Keller | d876c15 | 2017-07-10 13:23:10 -0700 | [diff] [blame] | 199 | for (vflre <<= 64 - i; vflre && i--; vflre += vflre) { |
| 200 | struct fm10k_vf_info *vf_info = &iov_data->vf_info[i]; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 201 | |
Jacob Keller | d876c15 | 2017-07-10 13:23:10 -0700 | [diff] [blame] | 202 | if (vflre >= 0) |
| 203 | continue; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 204 | |
Jacob Keller | d876c15 | 2017-07-10 13:23:10 -0700 | [diff] [blame] | 205 | hw->iov.ops.reset_resources(hw, vf_info); |
| 206 | vf_info->mbx.ops.connect(hw, &vf_info->mbx); |
| 207 | } |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 208 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 209 | read_unlock: |
| 210 | rcu_read_unlock(); |
| 211 | |
| 212 | return 0; |
| 213 | } |
| 214 | |
| 215 | s32 fm10k_iov_mbx(struct fm10k_intfc *interface) |
| 216 | { |
| 217 | struct fm10k_hw *hw = &interface->hw; |
| 218 | struct fm10k_iov_data *iov_data; |
| 219 | int i; |
| 220 | |
Jacob Keller | d8ec92f | 2016-02-10 14:45:51 -0800 | [diff] [blame] | 221 | /* if there is no iov_data then there is no mailbox to process */ |
Jacob Keller | ce4dad2 | 2016-06-17 16:21:11 -0700 | [diff] [blame] | 222 | if (!READ_ONCE(interface->iov_data)) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 223 | return 0; |
| 224 | |
| 225 | rcu_read_lock(); |
| 226 | |
| 227 | iov_data = interface->iov_data; |
| 228 | |
| 229 | /* check again now that we are in the RCU block */ |
| 230 | if (!iov_data) |
| 231 | goto read_unlock; |
| 232 | |
| 233 | /* lock the mailbox for transmit and receive */ |
| 234 | fm10k_mbx_lock(interface); |
| 235 | |
Jeff Kirsher | ada2411 | 2015-04-03 13:27:07 -0700 | [diff] [blame] | 236 | /* Most VF messages sent to the PF cause the PF to respond by |
| 237 | * requesting from the SM mailbox. This means that too many VF |
| 238 | * messages processed at once could cause a mailbox timeout on the PF. |
| 239 | * To prevent this, store a pointer to the next VF mbx to process. Use |
| 240 | * that as the start of the loop so that we don't starve whichever VF |
| 241 | * got ignored on the previous run. |
| 242 | */ |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 243 | process_mbx: |
| 244 | for (i = iov_data->next_vf_mbx ? : iov_data->num_vfs; i--;) { |
| 245 | struct fm10k_vf_info *vf_info = &iov_data->vf_info[i]; |
| 246 | struct fm10k_mbx_info *mbx = &vf_info->mbx; |
| 247 | u16 glort = vf_info->glort; |
| 248 | |
Jacob Keller | 17a9180 | 2017-10-02 07:17:50 -0700 | [diff] [blame] | 249 | /* process the SM mailbox first to drain outgoing messages */ |
| 250 | hw->mbx.ops.process(hw, &hw->mbx); |
| 251 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 252 | /* verify port mapping is valid, if not reset port */ |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 253 | if (vf_info->vf_flags && !fm10k_glort_valid_pf(hw, glort)) { |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 254 | hw->iov.ops.reset_lport(hw, vf_info); |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 255 | fm10k_clear_macvlan_queue(interface, glort, false); |
| 256 | } |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 257 | |
| 258 | /* reset VFs that have mailbox timed out */ |
| 259 | if (!mbx->timeout) { |
| 260 | hw->iov.ops.reset_resources(hw, vf_info); |
| 261 | mbx->ops.connect(hw, mbx); |
| 262 | } |
| 263 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 264 | /* guarantee we have free space in the SM mailbox */ |
Jacob Keller | 80043f3 | 2015-07-01 17:38:36 -0700 | [diff] [blame] | 265 | if (!hw->mbx.ops.tx_ready(&hw->mbx, FM10K_VFMBX_MSG_MTU)) { |
| 266 | /* keep track of how many times this occurs */ |
| 267 | interface->hw_sm_mbx_full++; |
Jacob Keller | b52b7f70 | 2017-03-08 15:55:43 -0800 | [diff] [blame] | 268 | |
| 269 | /* make sure we try again momentarily */ |
| 270 | fm10k_service_event_schedule(interface); |
| 271 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 272 | break; |
Jacob Keller | 80043f3 | 2015-07-01 17:38:36 -0700 | [diff] [blame] | 273 | } |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 274 | |
| 275 | /* cleanup mailbox and process received messages */ |
| 276 | mbx->ops.process(hw, mbx); |
| 277 | } |
| 278 | |
Jeff Kirsher | ada2411 | 2015-04-03 13:27:07 -0700 | [diff] [blame] | 279 | /* if we stopped processing mailboxes early, update next_vf_mbx. |
| 280 | * Otherwise, reset next_vf_mbx, and restart loop so that we process |
| 281 | * the remaining mailboxes we skipped at the start. |
| 282 | */ |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 283 | if (i >= 0) { |
| 284 | iov_data->next_vf_mbx = i + 1; |
| 285 | } else if (iov_data->next_vf_mbx) { |
| 286 | iov_data->next_vf_mbx = 0; |
| 287 | goto process_mbx; |
| 288 | } |
| 289 | |
| 290 | /* free the lock */ |
| 291 | fm10k_mbx_unlock(interface); |
| 292 | |
| 293 | read_unlock: |
| 294 | rcu_read_unlock(); |
| 295 | |
| 296 | return 0; |
| 297 | } |
| 298 | |
| 299 | void fm10k_iov_suspend(struct pci_dev *pdev) |
| 300 | { |
| 301 | struct fm10k_intfc *interface = pci_get_drvdata(pdev); |
| 302 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 303 | struct fm10k_hw *hw = &interface->hw; |
| 304 | int num_vfs, i; |
| 305 | |
| 306 | /* pull out num_vfs from iov_data */ |
| 307 | num_vfs = iov_data ? iov_data->num_vfs : 0; |
| 308 | |
| 309 | /* shut down queue mapping for VFs */ |
| 310 | fm10k_write_reg(hw, FM10K_DGLORTMAP(fm10k_dglort_vf_rss), |
| 311 | FM10K_DGLORTMAP_NONE); |
| 312 | |
| 313 | /* Stop any active VFs and reset their resources */ |
| 314 | for (i = 0; i < num_vfs; i++) { |
| 315 | struct fm10k_vf_info *vf_info = &iov_data->vf_info[i]; |
| 316 | |
| 317 | hw->iov.ops.reset_resources(hw, vf_info); |
| 318 | hw->iov.ops.reset_lport(hw, vf_info); |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 319 | fm10k_clear_macvlan_queue(interface, vf_info->glort, false); |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 320 | } |
| 321 | } |
| 322 | |
| 323 | int fm10k_iov_resume(struct pci_dev *pdev) |
| 324 | { |
| 325 | struct fm10k_intfc *interface = pci_get_drvdata(pdev); |
| 326 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 327 | struct fm10k_dglort_cfg dglort = { 0 }; |
| 328 | struct fm10k_hw *hw = &interface->hw; |
| 329 | int num_vfs, i; |
| 330 | |
| 331 | /* pull out num_vfs from iov_data */ |
| 332 | num_vfs = iov_data ? iov_data->num_vfs : 0; |
| 333 | |
| 334 | /* return error if iov_data is not already populated */ |
| 335 | if (!iov_data) |
| 336 | return -ENOMEM; |
| 337 | |
| 338 | /* allocate hardware resources for the VFs */ |
| 339 | hw->iov.ops.assign_resources(hw, num_vfs, num_vfs); |
| 340 | |
| 341 | /* configure DGLORT mapping for RSS */ |
| 342 | dglort.glort = hw->mac.dglort_map & FM10K_DGLORTMAP_NONE; |
| 343 | dglort.idx = fm10k_dglort_vf_rss; |
| 344 | dglort.inner_rss = 1; |
| 345 | dglort.rss_l = fls(fm10k_queues_per_pool(hw) - 1); |
| 346 | dglort.queue_b = fm10k_vf_queue_index(hw, 0); |
| 347 | dglort.vsi_l = fls(hw->iov.total_vfs - 1); |
| 348 | dglort.vsi_b = 1; |
| 349 | |
| 350 | hw->mac.ops.configure_dglort_map(hw, &dglort); |
| 351 | |
| 352 | /* assign resources to the device */ |
| 353 | for (i = 0; i < num_vfs; i++) { |
| 354 | struct fm10k_vf_info *vf_info = &iov_data->vf_info[i]; |
| 355 | |
| 356 | /* allocate all but the last GLORT to the VFs */ |
Jacob Keller | c8eeacb | 2018-01-24 14:17:15 -0800 | [diff] [blame] | 357 | if (i == (~hw->mac.dglort_map >> FM10K_DGLORTMAP_MASK_SHIFT)) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 358 | break; |
| 359 | |
| 360 | /* assign GLORT to VF, and restrict it to multicast */ |
| 361 | hw->iov.ops.set_lport(hw, vf_info, i, |
| 362 | FM10K_VF_FLAG_MULTI_CAPABLE); |
| 363 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 364 | /* mailbox is disconnected so we don't send a message */ |
| 365 | hw->iov.ops.assign_default_mac_vlan(hw, vf_info); |
| 366 | |
| 367 | /* now we are ready so we can connect */ |
| 368 | vf_info->mbx.ops.connect(hw, &vf_info->mbx); |
| 369 | } |
| 370 | |
| 371 | return 0; |
| 372 | } |
| 373 | |
| 374 | s32 fm10k_iov_update_pvid(struct fm10k_intfc *interface, u16 glort, u16 pvid) |
| 375 | { |
| 376 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 377 | struct fm10k_hw *hw = &interface->hw; |
| 378 | struct fm10k_vf_info *vf_info; |
| 379 | u16 vf_idx = (glort - hw->mac.dglort_map) & FM10K_DGLORTMAP_NONE; |
| 380 | |
| 381 | /* no IOV support, not our message to process */ |
| 382 | if (!iov_data) |
| 383 | return FM10K_ERR_PARAM; |
| 384 | |
| 385 | /* glort outside our range, not our message to process */ |
| 386 | if (vf_idx >= iov_data->num_vfs) |
| 387 | return FM10K_ERR_PARAM; |
| 388 | |
Matthew Vick | eca3204 | 2015-01-31 02:23:05 +0000 | [diff] [blame] | 389 | /* determine if an update has occurred and if so notify the VF */ |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 390 | vf_info = &iov_data->vf_info[vf_idx]; |
| 391 | if (vf_info->sw_vid != pvid) { |
| 392 | vf_info->sw_vid = pvid; |
| 393 | hw->iov.ops.assign_default_mac_vlan(hw, vf_info); |
| 394 | } |
| 395 | |
| 396 | return 0; |
| 397 | } |
| 398 | |
| 399 | static void fm10k_iov_free_data(struct pci_dev *pdev) |
| 400 | { |
| 401 | struct fm10k_intfc *interface = pci_get_drvdata(pdev); |
| 402 | |
| 403 | if (!interface->iov_data) |
| 404 | return; |
| 405 | |
| 406 | /* reclaim hardware resources */ |
| 407 | fm10k_iov_suspend(pdev); |
| 408 | |
| 409 | /* drop iov_data from interface */ |
| 410 | kfree_rcu(interface->iov_data, rcu); |
| 411 | interface->iov_data = NULL; |
| 412 | } |
| 413 | |
| 414 | static s32 fm10k_iov_alloc_data(struct pci_dev *pdev, int num_vfs) |
| 415 | { |
| 416 | struct fm10k_intfc *interface = pci_get_drvdata(pdev); |
| 417 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 418 | struct fm10k_hw *hw = &interface->hw; |
| 419 | size_t size; |
| 420 | int i, err; |
| 421 | |
| 422 | /* return error if iov_data is already populated */ |
| 423 | if (iov_data) |
| 424 | return -EBUSY; |
| 425 | |
| 426 | /* The PF should always be able to assign resources */ |
| 427 | if (!hw->iov.ops.assign_resources) |
| 428 | return -ENODEV; |
| 429 | |
| 430 | /* nothing to do if no VFs are requested */ |
| 431 | if (!num_vfs) |
| 432 | return 0; |
| 433 | |
| 434 | /* allocate memory for VF storage */ |
| 435 | size = offsetof(struct fm10k_iov_data, vf_info[num_vfs]); |
| 436 | iov_data = kzalloc(size, GFP_KERNEL); |
| 437 | if (!iov_data) |
| 438 | return -ENOMEM; |
| 439 | |
| 440 | /* record number of VFs */ |
| 441 | iov_data->num_vfs = num_vfs; |
| 442 | |
| 443 | /* loop through vf_info structures initializing each entry */ |
| 444 | for (i = 0; i < num_vfs; i++) { |
| 445 | struct fm10k_vf_info *vf_info = &iov_data->vf_info[i]; |
| 446 | |
| 447 | /* Record VF VSI value */ |
| 448 | vf_info->vsi = i + 1; |
| 449 | vf_info->vf_idx = i; |
| 450 | |
| 451 | /* initialize mailbox memory */ |
| 452 | err = fm10k_pfvf_mbx_init(hw, &vf_info->mbx, iov_mbx_data, i); |
| 453 | if (err) { |
| 454 | dev_err(&pdev->dev, |
| 455 | "Unable to initialize SR-IOV mailbox\n"); |
| 456 | kfree(iov_data); |
| 457 | return err; |
| 458 | } |
| 459 | } |
| 460 | |
| 461 | /* assign iov_data to interface */ |
| 462 | interface->iov_data = iov_data; |
| 463 | |
| 464 | /* allocate hardware resources for the VFs */ |
| 465 | fm10k_iov_resume(pdev); |
| 466 | |
| 467 | return 0; |
| 468 | } |
| 469 | |
| 470 | void fm10k_iov_disable(struct pci_dev *pdev) |
| 471 | { |
| 472 | if (pci_num_vf(pdev) && pci_vfs_assigned(pdev)) |
| 473 | dev_err(&pdev->dev, |
| 474 | "Cannot disable SR-IOV while VFs are assigned\n"); |
| 475 | else |
| 476 | pci_disable_sriov(pdev); |
| 477 | |
| 478 | fm10k_iov_free_data(pdev); |
| 479 | } |
| 480 | |
| 481 | static void fm10k_disable_aer_comp_abort(struct pci_dev *pdev) |
| 482 | { |
| 483 | u32 err_sev; |
| 484 | int pos; |
| 485 | |
| 486 | pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_ERR); |
| 487 | if (!pos) |
| 488 | return; |
| 489 | |
| 490 | pci_read_config_dword(pdev, pos + PCI_ERR_UNCOR_SEVER, &err_sev); |
| 491 | err_sev &= ~PCI_ERR_UNC_COMP_ABORT; |
| 492 | pci_write_config_dword(pdev, pos + PCI_ERR_UNCOR_SEVER, err_sev); |
| 493 | } |
| 494 | |
| 495 | int fm10k_iov_configure(struct pci_dev *pdev, int num_vfs) |
| 496 | { |
| 497 | int current_vfs = pci_num_vf(pdev); |
| 498 | int err = 0; |
| 499 | |
| 500 | if (current_vfs && pci_vfs_assigned(pdev)) { |
| 501 | dev_err(&pdev->dev, |
| 502 | "Cannot modify SR-IOV while VFs are assigned\n"); |
| 503 | num_vfs = current_vfs; |
| 504 | } else { |
| 505 | pci_disable_sriov(pdev); |
| 506 | fm10k_iov_free_data(pdev); |
| 507 | } |
| 508 | |
| 509 | /* allocate resources for the VFs */ |
| 510 | err = fm10k_iov_alloc_data(pdev, num_vfs); |
| 511 | if (err) |
| 512 | return err; |
| 513 | |
| 514 | /* allocate VFs if not already allocated */ |
Jacob Keller | c8eeacb | 2018-01-24 14:17:15 -0800 | [diff] [blame] | 515 | if (num_vfs && num_vfs != current_vfs) { |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 516 | /* Disable completer abort error reporting as |
| 517 | * the VFs can trigger this any time they read a queue |
| 518 | * that they don't own. |
| 519 | */ |
| 520 | fm10k_disable_aer_comp_abort(pdev); |
| 521 | |
| 522 | err = pci_enable_sriov(pdev, num_vfs); |
| 523 | if (err) { |
| 524 | dev_err(&pdev->dev, |
| 525 | "Enable PCI SR-IOV failed: %d\n", err); |
| 526 | return err; |
| 527 | } |
| 528 | } |
| 529 | |
| 530 | return num_vfs; |
| 531 | } |
| 532 | |
Jacob Keller | a38488f | 2015-06-03 16:31:07 -0700 | [diff] [blame] | 533 | static inline void fm10k_reset_vf_info(struct fm10k_intfc *interface, |
| 534 | struct fm10k_vf_info *vf_info) |
| 535 | { |
| 536 | struct fm10k_hw *hw = &interface->hw; |
| 537 | |
| 538 | /* assigning the MAC address will send a mailbox message */ |
| 539 | fm10k_mbx_lock(interface); |
| 540 | |
| 541 | /* disable LPORT for this VF which clears switch rules */ |
| 542 | hw->iov.ops.reset_lport(hw, vf_info); |
| 543 | |
Jacob Keller | 1f5c27e | 2017-07-10 13:23:18 -0700 | [diff] [blame] | 544 | fm10k_clear_macvlan_queue(interface, vf_info->glort, false); |
| 545 | |
Jacob Keller | a38488f | 2015-06-03 16:31:07 -0700 | [diff] [blame] | 546 | /* assign new MAC+VLAN for this VF */ |
| 547 | hw->iov.ops.assign_default_mac_vlan(hw, vf_info); |
| 548 | |
| 549 | /* re-enable the LPORT for this VF */ |
| 550 | hw->iov.ops.set_lport(hw, vf_info, vf_info->vf_idx, |
| 551 | FM10K_VF_FLAG_MULTI_CAPABLE); |
| 552 | |
| 553 | fm10k_mbx_unlock(interface); |
| 554 | } |
| 555 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 556 | int fm10k_ndo_set_vf_mac(struct net_device *netdev, int vf_idx, u8 *mac) |
| 557 | { |
| 558 | struct fm10k_intfc *interface = netdev_priv(netdev); |
| 559 | struct fm10k_iov_data *iov_data = interface->iov_data; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 560 | struct fm10k_vf_info *vf_info; |
| 561 | |
| 562 | /* verify SR-IOV is active and that vf idx is valid */ |
| 563 | if (!iov_data || vf_idx >= iov_data->num_vfs) |
| 564 | return -EINVAL; |
| 565 | |
| 566 | /* verify MAC addr is valid */ |
| 567 | if (!is_zero_ether_addr(mac) && !is_valid_ether_addr(mac)) |
| 568 | return -EINVAL; |
| 569 | |
| 570 | /* record new MAC address */ |
| 571 | vf_info = &iov_data->vf_info[vf_idx]; |
| 572 | ether_addr_copy(vf_info->mac, mac); |
| 573 | |
Jacob Keller | a38488f | 2015-06-03 16:31:07 -0700 | [diff] [blame] | 574 | fm10k_reset_vf_info(interface, vf_info); |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 575 | |
| 576 | return 0; |
| 577 | } |
| 578 | |
| 579 | int fm10k_ndo_set_vf_vlan(struct net_device *netdev, int vf_idx, u16 vid, |
Moshe Shemesh | 79aab09 | 2016-09-22 12:11:15 +0300 | [diff] [blame] | 580 | u8 qos, __be16 vlan_proto) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 581 | { |
| 582 | struct fm10k_intfc *interface = netdev_priv(netdev); |
| 583 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 584 | struct fm10k_hw *hw = &interface->hw; |
| 585 | struct fm10k_vf_info *vf_info; |
| 586 | |
| 587 | /* verify SR-IOV is active and that vf idx is valid */ |
| 588 | if (!iov_data || vf_idx >= iov_data->num_vfs) |
| 589 | return -EINVAL; |
| 590 | |
| 591 | /* QOS is unsupported and VLAN IDs accepted range 0-4094 */ |
| 592 | if (qos || (vid > (VLAN_VID_MASK - 1))) |
| 593 | return -EINVAL; |
| 594 | |
Moshe Shemesh | 79aab09 | 2016-09-22 12:11:15 +0300 | [diff] [blame] | 595 | /* VF VLAN Protocol part to default is unsupported */ |
| 596 | if (vlan_proto != htons(ETH_P_8021Q)) |
| 597 | return -EPROTONOSUPPORT; |
| 598 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 599 | vf_info = &iov_data->vf_info[vf_idx]; |
| 600 | |
| 601 | /* exit if there is nothing to do */ |
| 602 | if (vf_info->pf_vid == vid) |
| 603 | return 0; |
| 604 | |
| 605 | /* record default VLAN ID for VF */ |
| 606 | vf_info->pf_vid = vid; |
| 607 | |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 608 | /* Clear the VLAN table for the VF */ |
| 609 | hw->mac.ops.update_vlan(hw, FM10K_VLAN_ALL, vf_info->vsi, false); |
| 610 | |
Jacob Keller | a38488f | 2015-06-03 16:31:07 -0700 | [diff] [blame] | 611 | fm10k_reset_vf_info(interface, vf_info); |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 612 | |
| 613 | return 0; |
| 614 | } |
| 615 | |
Jeff Kirsher | de44519 | 2015-04-03 13:26:56 -0700 | [diff] [blame] | 616 | int fm10k_ndo_set_vf_bw(struct net_device *netdev, int vf_idx, |
Jacob Keller | 3e256ac | 2017-08-11 11:14:58 -0700 | [diff] [blame] | 617 | int __always_unused min_rate, int max_rate) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 618 | { |
| 619 | struct fm10k_intfc *interface = netdev_priv(netdev); |
| 620 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 621 | struct fm10k_hw *hw = &interface->hw; |
| 622 | |
| 623 | /* verify SR-IOV is active and that vf idx is valid */ |
| 624 | if (!iov_data || vf_idx >= iov_data->num_vfs) |
| 625 | return -EINVAL; |
| 626 | |
| 627 | /* rate limit cannot be less than 10Mbs or greater than link speed */ |
Jacob Keller | 3e256ac | 2017-08-11 11:14:58 -0700 | [diff] [blame] | 628 | if (max_rate && |
| 629 | (max_rate < FM10K_VF_TC_MIN || max_rate > FM10K_VF_TC_MAX)) |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 630 | return -EINVAL; |
| 631 | |
| 632 | /* store values */ |
Jacob Keller | 3e256ac | 2017-08-11 11:14:58 -0700 | [diff] [blame] | 633 | iov_data->vf_info[vf_idx].rate = max_rate; |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 634 | |
| 635 | /* update hardware configuration */ |
Jacob Keller | 3e256ac | 2017-08-11 11:14:58 -0700 | [diff] [blame] | 636 | hw->iov.ops.configure_tc(hw, vf_idx, max_rate); |
Alexander Duyck | 883a9cc | 2014-09-20 19:52:09 -0400 | [diff] [blame] | 637 | |
| 638 | return 0; |
| 639 | } |
| 640 | |
| 641 | int fm10k_ndo_get_vf_config(struct net_device *netdev, |
| 642 | int vf_idx, struct ifla_vf_info *ivi) |
| 643 | { |
| 644 | struct fm10k_intfc *interface = netdev_priv(netdev); |
| 645 | struct fm10k_iov_data *iov_data = interface->iov_data; |
| 646 | struct fm10k_vf_info *vf_info; |
| 647 | |
| 648 | /* verify SR-IOV is active and that vf idx is valid */ |
| 649 | if (!iov_data || vf_idx >= iov_data->num_vfs) |
| 650 | return -EINVAL; |
| 651 | |
| 652 | vf_info = &iov_data->vf_info[vf_idx]; |
| 653 | |
| 654 | ivi->vf = vf_idx; |
| 655 | ivi->max_tx_rate = vf_info->rate; |
| 656 | ivi->min_tx_rate = 0; |
| 657 | ether_addr_copy(ivi->mac, vf_info->mac); |
| 658 | ivi->vlan = vf_info->pf_vid; |
| 659 | ivi->qos = 0; |
| 660 | |
| 661 | return 0; |
| 662 | } |