1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright(c) 2013 - 2019 Intel Corporation. */
8 static s32
fm10k_iov_msg_error(struct fm10k_hw
*hw
, u32
**results
,
9 struct fm10k_mbx_info
*mbx
)
11 struct fm10k_vf_info
*vf_info
= (struct fm10k_vf_info
*)mbx
;
12 struct fm10k_intfc
*interface
= hw
->back
;
13 struct pci_dev
*pdev
= interface
->pdev
;
15 dev_err(&pdev
->dev
, "Unknown message ID %u on VF %d\n",
16 **results
& FM10K_TLV_ID_MASK
, vf_info
->vf_idx
);
18 return fm10k_tlv_msg_error(hw
, results
, mbx
);
22 * fm10k_iov_msg_queue_mac_vlan - Message handler for MAC/VLAN request from VF
23 * @hw: Pointer to hardware structure
24 * @results: Pointer array to message, results[0] is pointer to message
25 * @mbx: Pointer to mailbox information structure
27 * This function is a custom handler for MAC/VLAN requests from the VF. The
28 * assumption is that it is acceptable to directly hand off the message from
29 * the VF to the PF's switch manager. However, we use a MAC/VLAN message
30 * queue to avoid overloading the mailbox when a large number of requests
33 static s32
fm10k_iov_msg_queue_mac_vlan(struct fm10k_hw
*hw
, u32
**results
,
34 struct fm10k_mbx_info
*mbx
)
36 struct fm10k_vf_info
*vf_info
= (struct fm10k_vf_info
*)mbx
;
37 struct fm10k_intfc
*interface
= hw
->back
;
45 /* we shouldn't be updating rules on a disabled interface */
46 if (!FM10K_VF_FLAG_ENABLED(vf_info
))
47 err
= FM10K_ERR_PARAM
;
49 if (!err
&& !!results
[FM10K_MAC_VLAN_MSG_VLAN
]) {
50 result
= results
[FM10K_MAC_VLAN_MSG_VLAN
];
52 /* record VLAN id requested */
53 err
= fm10k_tlv_attr_get_u32(result
, &vid
);
57 set
= !(vid
& FM10K_VLAN_CLEAR
);
58 vid
&= ~FM10K_VLAN_CLEAR
;
60 /* if the length field has been set, this is a multi-bit
61 * update request. For multi-bit requests, simply disallow
62 * them when the pf_vid has been set. In this case, the PF
63 * should have already cleared the VLAN_TABLE, and if we
64 * allowed them, it could allow a rogue VF to receive traffic
65 * on a VLAN it was not assigned. In the single-bit case, we
66 * need to modify requests for VLAN 0 to use the default PF or
67 * SW vid when assigned.
71 /* prevent multi-bit requests when PF has
72 * administratively set the VLAN for this VF
75 return FM10K_ERR_PARAM
;
77 err
= fm10k_iov_select_vid(vf_info
, (u16
)vid
);
84 /* update VSI info for VF in regards to VLAN table */
85 err
= hw
->mac
.ops
.update_vlan(hw
, vid
, vf_info
->vsi
, set
);
88 if (!err
&& !!results
[FM10K_MAC_VLAN_MSG_MAC
]) {
89 result
= results
[FM10K_MAC_VLAN_MSG_MAC
];
91 /* record unicast MAC address requested */
92 err
= fm10k_tlv_attr_get_mac_vlan(result
, mac
, &vlan
);
96 /* block attempts to set MAC for a locked device */
97 if (is_valid_ether_addr(vf_info
->mac
) &&
98 !ether_addr_equal(mac
, vf_info
->mac
))
99 return FM10K_ERR_PARAM
;
101 set
= !(vlan
& FM10K_VLAN_CLEAR
);
102 vlan
&= ~FM10K_VLAN_CLEAR
;
104 err
= fm10k_iov_select_vid(vf_info
, vlan
);
110 /* Add this request to the MAC/VLAN queue */
111 err
= fm10k_queue_mac_request(interface
, vf_info
->glort
,
115 if (!err
&& !!results
[FM10K_MAC_VLAN_MSG_MULTICAST
]) {
116 result
= results
[FM10K_MAC_VLAN_MSG_MULTICAST
];
118 /* record multicast MAC address requested */
119 err
= fm10k_tlv_attr_get_mac_vlan(result
, mac
, &vlan
);
123 /* verify that the VF is allowed to request multicast */
124 if (!(vf_info
->vf_flags
& FM10K_VF_FLAG_MULTI_ENABLED
))
125 return FM10K_ERR_PARAM
;
127 set
= !(vlan
& FM10K_VLAN_CLEAR
);
128 vlan
&= ~FM10K_VLAN_CLEAR
;
130 err
= fm10k_iov_select_vid(vf_info
, vlan
);
136 /* Add this request to the MAC/VLAN queue */
137 err
= fm10k_queue_mac_request(interface
, vf_info
->glort
,
144 static const struct fm10k_msg_data iov_mbx_data
[] = {
145 FM10K_TLV_MSG_TEST_HANDLER(fm10k_tlv_msg_test
),
146 FM10K_VF_MSG_MSIX_HANDLER(fm10k_iov_msg_msix_pf
),
147 FM10K_VF_MSG_MAC_VLAN_HANDLER(fm10k_iov_msg_queue_mac_vlan
),
148 FM10K_VF_MSG_LPORT_STATE_HANDLER(fm10k_iov_msg_lport_state_pf
),
149 FM10K_TLV_MSG_ERROR_HANDLER(fm10k_iov_msg_error
),
152 s32
fm10k_iov_event(struct fm10k_intfc
*interface
)
154 struct fm10k_hw
*hw
= &interface
->hw
;
155 struct fm10k_iov_data
*iov_data
;
159 /* if there is no iov_data then there is no mailbox to process */
160 if (!READ_ONCE(interface
->iov_data
))
165 iov_data
= interface
->iov_data
;
167 /* check again now that we are in the RCU block */
171 if (!(fm10k_read_reg(hw
, FM10K_EICR
) & FM10K_EICR_VFLR
))
174 /* read VFLRE to determine if any VFs have been reset */
175 vflre
= fm10k_read_reg(hw
, FM10K_PFVFLRE(1));
177 vflre
|= fm10k_read_reg(hw
, FM10K_PFVFLRE(0));
179 i
= iov_data
->num_vfs
;
181 for (vflre
<<= 64 - i
; vflre
&& i
--; vflre
+= vflre
) {
182 struct fm10k_vf_info
*vf_info
= &iov_data
->vf_info
[i
];
187 hw
->iov
.ops
.reset_resources(hw
, vf_info
);
188 vf_info
->mbx
.ops
.connect(hw
, &vf_info
->mbx
);
197 s32
fm10k_iov_mbx(struct fm10k_intfc
*interface
)
199 struct fm10k_hw
*hw
= &interface
->hw
;
200 struct fm10k_iov_data
*iov_data
;
203 /* if there is no iov_data then there is no mailbox to process */
204 if (!READ_ONCE(interface
->iov_data
))
209 iov_data
= interface
->iov_data
;
211 /* check again now that we are in the RCU block */
215 /* lock the mailbox for transmit and receive */
216 fm10k_mbx_lock(interface
);
218 /* Most VF messages sent to the PF cause the PF to respond by
219 * requesting from the SM mailbox. This means that too many VF
220 * messages processed at once could cause a mailbox timeout on the PF.
221 * To prevent this, store a pointer to the next VF mbx to process. Use
222 * that as the start of the loop so that we don't starve whichever VF
223 * got ignored on the previous run.
226 for (i
= iov_data
->next_vf_mbx
? : iov_data
->num_vfs
; i
--;) {
227 struct fm10k_vf_info
*vf_info
= &iov_data
->vf_info
[i
];
228 struct fm10k_mbx_info
*mbx
= &vf_info
->mbx
;
229 u16 glort
= vf_info
->glort
;
231 /* process the SM mailbox first to drain outgoing messages */
232 hw
->mbx
.ops
.process(hw
, &hw
->mbx
);
234 /* verify port mapping is valid, if not reset port */
235 if (vf_info
->vf_flags
&& !fm10k_glort_valid_pf(hw
, glort
)) {
236 hw
->iov
.ops
.reset_lport(hw
, vf_info
);
237 fm10k_clear_macvlan_queue(interface
, glort
, false);
240 /* reset VFs that have mailbox timed out */
242 hw
->iov
.ops
.reset_resources(hw
, vf_info
);
243 mbx
->ops
.connect(hw
, mbx
);
246 /* guarantee we have free space in the SM mailbox */
247 if (hw
->mbx
.state
== FM10K_STATE_OPEN
&&
248 !hw
->mbx
.ops
.tx_ready(&hw
->mbx
, FM10K_VFMBX_MSG_MTU
)) {
249 /* keep track of how many times this occurs */
250 interface
->hw_sm_mbx_full
++;
252 /* make sure we try again momentarily */
253 fm10k_service_event_schedule(interface
);
258 /* cleanup mailbox and process received messages */
259 mbx
->ops
.process(hw
, mbx
);
262 /* if we stopped processing mailboxes early, update next_vf_mbx.
263 * Otherwise, reset next_vf_mbx, and restart loop so that we process
264 * the remaining mailboxes we skipped at the start.
267 iov_data
->next_vf_mbx
= i
+ 1;
268 } else if (iov_data
->next_vf_mbx
) {
269 iov_data
->next_vf_mbx
= 0;
274 fm10k_mbx_unlock(interface
);
282 void fm10k_iov_suspend(struct pci_dev
*pdev
)
284 struct fm10k_intfc
*interface
= pci_get_drvdata(pdev
);
285 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
286 struct fm10k_hw
*hw
= &interface
->hw
;
289 /* pull out num_vfs from iov_data */
290 num_vfs
= iov_data
? iov_data
->num_vfs
: 0;
292 /* shut down queue mapping for VFs */
293 fm10k_write_reg(hw
, FM10K_DGLORTMAP(fm10k_dglort_vf_rss
),
294 FM10K_DGLORTMAP_NONE
);
296 /* Stop any active VFs and reset their resources */
297 for (i
= 0; i
< num_vfs
; i
++) {
298 struct fm10k_vf_info
*vf_info
= &iov_data
->vf_info
[i
];
300 hw
->iov
.ops
.reset_resources(hw
, vf_info
);
301 hw
->iov
.ops
.reset_lport(hw
, vf_info
);
302 fm10k_clear_macvlan_queue(interface
, vf_info
->glort
, false);
306 static void fm10k_mask_aer_comp_abort(struct pci_dev
*pdev
)
311 pos
= pci_find_ext_capability(pdev
, PCI_EXT_CAP_ID_ERR
);
315 /* Mask the completion abort bit in the ERR_UNCOR_MASK register,
316 * preventing the device from reporting these errors to the upstream
317 * PCIe root device. This avoids bringing down platforms which upgrade
318 * non-fatal completer aborts into machine check exceptions. Completer
319 * aborts can occur whenever a VF reads a queue it doesn't own.
321 pci_read_config_dword(pdev
, pos
+ PCI_ERR_UNCOR_MASK
, &err_mask
);
322 err_mask
|= PCI_ERR_UNC_COMP_ABORT
;
323 pci_write_config_dword(pdev
, pos
+ PCI_ERR_UNCOR_MASK
, err_mask
);
326 int fm10k_iov_resume(struct pci_dev
*pdev
)
328 struct fm10k_intfc
*interface
= pci_get_drvdata(pdev
);
329 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
330 struct fm10k_dglort_cfg dglort
= { 0 };
331 struct fm10k_hw
*hw
= &interface
->hw
;
334 /* pull out num_vfs from iov_data */
335 num_vfs
= iov_data
? iov_data
->num_vfs
: 0;
337 /* return error if iov_data is not already populated */
341 /* Lower severity of completer abort error reporting as
342 * the VFs can trigger this any time they read a queue
343 * that they don't own.
345 fm10k_mask_aer_comp_abort(pdev
);
347 /* allocate hardware resources for the VFs */
348 hw
->iov
.ops
.assign_resources(hw
, num_vfs
, num_vfs
);
350 /* configure DGLORT mapping for RSS */
351 dglort
.glort
= hw
->mac
.dglort_map
& FM10K_DGLORTMAP_NONE
;
352 dglort
.idx
= fm10k_dglort_vf_rss
;
353 dglort
.inner_rss
= 1;
354 dglort
.rss_l
= fls(fm10k_queues_per_pool(hw
) - 1);
355 dglort
.queue_b
= fm10k_vf_queue_index(hw
, 0);
356 dglort
.vsi_l
= fls(hw
->iov
.total_vfs
- 1);
359 hw
->mac
.ops
.configure_dglort_map(hw
, &dglort
);
361 /* assign resources to the device */
362 for (i
= 0; i
< num_vfs
; i
++) {
363 struct fm10k_vf_info
*vf_info
= &iov_data
->vf_info
[i
];
365 /* allocate all but the last GLORT to the VFs */
366 if (i
== (~hw
->mac
.dglort_map
>> FM10K_DGLORTMAP_MASK_SHIFT
))
369 /* assign GLORT to VF, and restrict it to multicast */
370 hw
->iov
.ops
.set_lport(hw
, vf_info
, i
,
371 FM10K_VF_FLAG_MULTI_CAPABLE
);
373 /* mailbox is disconnected so we don't send a message */
374 hw
->iov
.ops
.assign_default_mac_vlan(hw
, vf_info
);
376 /* now we are ready so we can connect */
377 vf_info
->mbx
.ops
.connect(hw
, &vf_info
->mbx
);
383 s32
fm10k_iov_update_pvid(struct fm10k_intfc
*interface
, u16 glort
, u16 pvid
)
385 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
386 struct fm10k_hw
*hw
= &interface
->hw
;
387 struct fm10k_vf_info
*vf_info
;
388 u16 vf_idx
= (glort
- hw
->mac
.dglort_map
) & FM10K_DGLORTMAP_NONE
;
390 /* no IOV support, not our message to process */
392 return FM10K_ERR_PARAM
;
394 /* glort outside our range, not our message to process */
395 if (vf_idx
>= iov_data
->num_vfs
)
396 return FM10K_ERR_PARAM
;
398 /* determine if an update has occurred and if so notify the VF */
399 vf_info
= &iov_data
->vf_info
[vf_idx
];
400 if (vf_info
->sw_vid
!= pvid
) {
401 vf_info
->sw_vid
= pvid
;
402 hw
->iov
.ops
.assign_default_mac_vlan(hw
, vf_info
);
408 static void fm10k_iov_free_data(struct pci_dev
*pdev
)
410 struct fm10k_intfc
*interface
= pci_get_drvdata(pdev
);
412 if (!interface
->iov_data
)
415 /* reclaim hardware resources */
416 fm10k_iov_suspend(pdev
);
418 /* drop iov_data from interface */
419 kfree_rcu(interface
->iov_data
, rcu
);
420 interface
->iov_data
= NULL
;
423 static s32
fm10k_iov_alloc_data(struct pci_dev
*pdev
, int num_vfs
)
425 struct fm10k_intfc
*interface
= pci_get_drvdata(pdev
);
426 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
427 struct fm10k_hw
*hw
= &interface
->hw
;
431 /* return error if iov_data is already populated */
435 /* The PF should always be able to assign resources */
436 if (!hw
->iov
.ops
.assign_resources
)
439 /* nothing to do if no VFs are requested */
443 /* allocate memory for VF storage */
444 size
= offsetof(struct fm10k_iov_data
, vf_info
[num_vfs
]);
445 iov_data
= kzalloc(size
, GFP_KERNEL
);
449 /* record number of VFs */
450 iov_data
->num_vfs
= num_vfs
;
452 /* loop through vf_info structures initializing each entry */
453 for (i
= 0; i
< num_vfs
; i
++) {
454 struct fm10k_vf_info
*vf_info
= &iov_data
->vf_info
[i
];
457 /* Record VF VSI value */
458 vf_info
->vsi
= i
+ 1;
461 /* initialize mailbox memory */
462 err
= fm10k_pfvf_mbx_init(hw
, &vf_info
->mbx
, iov_mbx_data
, i
);
465 "Unable to initialize SR-IOV mailbox\n");
471 /* assign iov_data to interface */
472 interface
->iov_data
= iov_data
;
474 /* allocate hardware resources for the VFs */
475 fm10k_iov_resume(pdev
);
480 void fm10k_iov_disable(struct pci_dev
*pdev
)
482 if (pci_num_vf(pdev
) && pci_vfs_assigned(pdev
))
484 "Cannot disable SR-IOV while VFs are assigned\n");
486 pci_disable_sriov(pdev
);
488 fm10k_iov_free_data(pdev
);
491 int fm10k_iov_configure(struct pci_dev
*pdev
, int num_vfs
)
493 int current_vfs
= pci_num_vf(pdev
);
496 if (current_vfs
&& pci_vfs_assigned(pdev
)) {
498 "Cannot modify SR-IOV while VFs are assigned\n");
499 num_vfs
= current_vfs
;
501 pci_disable_sriov(pdev
);
502 fm10k_iov_free_data(pdev
);
505 /* allocate resources for the VFs */
506 err
= fm10k_iov_alloc_data(pdev
, num_vfs
);
510 /* allocate VFs if not already allocated */
511 if (num_vfs
&& num_vfs
!= current_vfs
) {
512 err
= pci_enable_sriov(pdev
, num_vfs
);
515 "Enable PCI SR-IOV failed: %d\n", err
);
524 * fm10k_iov_update_stats - Update stats for all VFs
525 * @interface: device private structure
527 * Updates the VF statistics for all enabled VFs. Expects to be called by
528 * fm10k_update_stats and assumes that locking via the __FM10K_UPDATING_STATS
529 * bit is already handled.
531 void fm10k_iov_update_stats(struct fm10k_intfc
*interface
)
533 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
534 struct fm10k_hw
*hw
= &interface
->hw
;
540 for (i
= 0; i
< iov_data
->num_vfs
; i
++)
541 hw
->iov
.ops
.update_stats(hw
, iov_data
->vf_info
[i
].stats
, i
);
544 static inline void fm10k_reset_vf_info(struct fm10k_intfc
*interface
,
545 struct fm10k_vf_info
*vf_info
)
547 struct fm10k_hw
*hw
= &interface
->hw
;
549 /* assigning the MAC address will send a mailbox message */
550 fm10k_mbx_lock(interface
);
552 /* disable LPORT for this VF which clears switch rules */
553 hw
->iov
.ops
.reset_lport(hw
, vf_info
);
555 fm10k_clear_macvlan_queue(interface
, vf_info
->glort
, false);
557 /* assign new MAC+VLAN for this VF */
558 hw
->iov
.ops
.assign_default_mac_vlan(hw
, vf_info
);
560 /* re-enable the LPORT for this VF */
561 hw
->iov
.ops
.set_lport(hw
, vf_info
, vf_info
->vf_idx
,
562 FM10K_VF_FLAG_MULTI_CAPABLE
);
564 fm10k_mbx_unlock(interface
);
567 int fm10k_ndo_set_vf_mac(struct net_device
*netdev
, int vf_idx
, u8
*mac
)
569 struct fm10k_intfc
*interface
= netdev_priv(netdev
);
570 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
571 struct fm10k_vf_info
*vf_info
;
573 /* verify SR-IOV is active and that vf idx is valid */
574 if (!iov_data
|| vf_idx
>= iov_data
->num_vfs
)
577 /* verify MAC addr is valid */
578 if (!is_zero_ether_addr(mac
) && !is_valid_ether_addr(mac
))
581 /* record new MAC address */
582 vf_info
= &iov_data
->vf_info
[vf_idx
];
583 ether_addr_copy(vf_info
->mac
, mac
);
585 fm10k_reset_vf_info(interface
, vf_info
);
590 int fm10k_ndo_set_vf_vlan(struct net_device
*netdev
, int vf_idx
, u16 vid
,
591 u8 qos
, __be16 vlan_proto
)
593 struct fm10k_intfc
*interface
= netdev_priv(netdev
);
594 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
595 struct fm10k_hw
*hw
= &interface
->hw
;
596 struct fm10k_vf_info
*vf_info
;
598 /* verify SR-IOV is active and that vf idx is valid */
599 if (!iov_data
|| vf_idx
>= iov_data
->num_vfs
)
602 /* QOS is unsupported and VLAN IDs accepted range 0-4094 */
603 if (qos
|| (vid
> (VLAN_VID_MASK
- 1)))
606 /* VF VLAN Protocol part to default is unsupported */
607 if (vlan_proto
!= htons(ETH_P_8021Q
))
608 return -EPROTONOSUPPORT
;
610 vf_info
= &iov_data
->vf_info
[vf_idx
];
612 /* exit if there is nothing to do */
613 if (vf_info
->pf_vid
== vid
)
616 /* record default VLAN ID for VF */
617 vf_info
->pf_vid
= vid
;
619 /* Clear the VLAN table for the VF */
620 hw
->mac
.ops
.update_vlan(hw
, FM10K_VLAN_ALL
, vf_info
->vsi
, false);
622 fm10k_reset_vf_info(interface
, vf_info
);
627 int fm10k_ndo_set_vf_bw(struct net_device
*netdev
, int vf_idx
,
628 int __always_unused min_rate
, int max_rate
)
630 struct fm10k_intfc
*interface
= netdev_priv(netdev
);
631 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
632 struct fm10k_hw
*hw
= &interface
->hw
;
634 /* verify SR-IOV is active and that vf idx is valid */
635 if (!iov_data
|| vf_idx
>= iov_data
->num_vfs
)
638 /* rate limit cannot be less than 10Mbs or greater than link speed */
640 (max_rate
< FM10K_VF_TC_MIN
|| max_rate
> FM10K_VF_TC_MAX
))
644 iov_data
->vf_info
[vf_idx
].rate
= max_rate
;
646 /* update hardware configuration */
647 hw
->iov
.ops
.configure_tc(hw
, vf_idx
, max_rate
);
652 int fm10k_ndo_get_vf_config(struct net_device
*netdev
,
653 int vf_idx
, struct ifla_vf_info
*ivi
)
655 struct fm10k_intfc
*interface
= netdev_priv(netdev
);
656 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
657 struct fm10k_vf_info
*vf_info
;
659 /* verify SR-IOV is active and that vf idx is valid */
660 if (!iov_data
|| vf_idx
>= iov_data
->num_vfs
)
663 vf_info
= &iov_data
->vf_info
[vf_idx
];
666 ivi
->max_tx_rate
= vf_info
->rate
;
667 ivi
->min_tx_rate
= 0;
668 ether_addr_copy(ivi
->mac
, vf_info
->mac
);
669 ivi
->vlan
= vf_info
->pf_vid
;
675 int fm10k_ndo_get_vf_stats(struct net_device
*netdev
,
676 int vf_idx
, struct ifla_vf_stats
*stats
)
678 struct fm10k_intfc
*interface
= netdev_priv(netdev
);
679 struct fm10k_iov_data
*iov_data
= interface
->iov_data
;
680 struct fm10k_hw
*hw
= &interface
->hw
;
681 struct fm10k_hw_stats_q
*hw_stats
;
684 /* verify SR-IOV is active and that vf idx is valid */
685 if (!iov_data
|| vf_idx
>= iov_data
->num_vfs
)
688 qpp
= fm10k_queues_per_pool(hw
);
689 hw_stats
= iov_data
->vf_info
[vf_idx
].stats
;
691 for (idx
= 0; idx
< qpp
; idx
++) {
692 stats
->rx_packets
+= hw_stats
[idx
].rx_packets
.count
;
693 stats
->tx_packets
+= hw_stats
[idx
].tx_packets
.count
;
694 stats
->rx_bytes
+= hw_stats
[idx
].rx_bytes
.count
;
695 stats
->tx_bytes
+= hw_stats
[idx
].tx_bytes
.count
;
696 stats
->rx_dropped
+= hw_stats
[idx
].rx_drops
.count
;