1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (C) 2022, Intel Corporation. */
3
4 #include "ice_vf_lib_private.h"
5 #include "ice.h"
6 #include "ice_lib.h"
7 #include "ice_fltr.h"
8 #include "ice_virtchnl_allowlist.h"
9
10 /* Public functions which may be accessed by all driver files */
11
12 /**
13 * ice_get_vf_by_id - Get pointer to VF by ID
14 * @pf: the PF private structure
15 * @vf_id: the VF ID to locate
16 *
17 * Locate and return a pointer to the VF structure associated with a given ID.
18 * Returns NULL if the ID does not have a valid VF structure associated with
19 * it.
20 *
21 * This function takes a reference to the VF, which must be released by
22 * calling ice_put_vf() once the caller is finished accessing the VF structure
23 * returned.
24 */
ice_get_vf_by_id(struct ice_pf * pf,u16 vf_id)25 struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id)
26 {
27 struct ice_vf *vf;
28
29 rcu_read_lock();
30 hash_for_each_possible_rcu(pf->vfs.table, vf, entry, vf_id) {
31 if (vf->vf_id == vf_id) {
32 struct ice_vf *found;
33
34 if (kref_get_unless_zero(&vf->refcnt))
35 found = vf;
36 else
37 found = NULL;
38
39 rcu_read_unlock();
40 return found;
41 }
42 }
43 rcu_read_unlock();
44
45 return NULL;
46 }
47
48 /**
49 * ice_release_vf - Release VF associated with a refcount
50 * @ref: the kref decremented to zero
51 *
52 * Callback function for kref_put to release a VF once its reference count has
53 * hit zero.
54 */
ice_release_vf(struct kref * ref)55 static void ice_release_vf(struct kref *ref)
56 {
57 struct ice_vf *vf = container_of(ref, struct ice_vf, refcnt);
58
59 vf->vf_ops->free(vf);
60 }
61
62 /**
63 * ice_put_vf - Release a reference to a VF
64 * @vf: the VF structure to decrease reference count on
65 *
66 * Decrease the reference count for a VF, and free the entry if it is no
67 * longer in use.
68 *
69 * This must be called after ice_get_vf_by_id() once the reference to the VF
70 * structure is no longer used. Otherwise, the VF structure will never be
71 * freed.
72 */
ice_put_vf(struct ice_vf * vf)73 void ice_put_vf(struct ice_vf *vf)
74 {
75 kref_put(&vf->refcnt, ice_release_vf);
76 }
77
78 /**
79 * ice_has_vfs - Return true if the PF has any associated VFs
80 * @pf: the PF private structure
81 *
82 * Return whether or not the PF has any allocated VFs.
83 *
84 * Note that this function only guarantees that there are no VFs at the point
85 * of calling it. It does not guarantee that no more VFs will be added.
86 */
ice_has_vfs(struct ice_pf * pf)87 bool ice_has_vfs(struct ice_pf *pf)
88 {
89 /* A simple check that the hash table is not empty does not require
90 * the mutex or rcu_read_lock.
91 */
92 return !hash_empty(pf->vfs.table);
93 }
94
95 /**
96 * ice_get_num_vfs - Get number of allocated VFs
97 * @pf: the PF private structure
98 *
99 * Return the total number of allocated VFs. NOTE: VF IDs are not guaranteed
100 * to be contiguous. Do not assume that a VF ID is guaranteed to be less than
101 * the output of this function.
102 */
ice_get_num_vfs(struct ice_pf * pf)103 u16 ice_get_num_vfs(struct ice_pf *pf)
104 {
105 struct ice_vf *vf;
106 unsigned int bkt;
107 u16 num_vfs = 0;
108
109 rcu_read_lock();
110 ice_for_each_vf_rcu(pf, bkt, vf)
111 num_vfs++;
112 rcu_read_unlock();
113
114 return num_vfs;
115 }
116
117 /**
118 * ice_get_vf_vsi - get VF's VSI based on the stored index
119 * @vf: VF used to get VSI
120 */
ice_get_vf_vsi(struct ice_vf * vf)121 struct ice_vsi *ice_get_vf_vsi(struct ice_vf *vf)
122 {
123 if (vf->lan_vsi_idx == ICE_NO_VSI)
124 return NULL;
125
126 return vf->pf->vsi[vf->lan_vsi_idx];
127 }
128
129 /**
130 * ice_is_vf_disabled
131 * @vf: pointer to the VF info
132 *
133 * If the PF has been disabled, there is no need resetting VF until PF is
134 * active again. Similarly, if the VF has been disabled, this means something
135 * else is resetting the VF, so we shouldn't continue.
136 *
137 * Returns true if the caller should consider the VF as disabled whether
138 * because that single VF is explicitly disabled or because the PF is
139 * currently disabled.
140 */
ice_is_vf_disabled(struct ice_vf * vf)141 bool ice_is_vf_disabled(struct ice_vf *vf)
142 {
143 struct ice_pf *pf = vf->pf;
144
145 return (test_bit(ICE_VF_DIS, pf->state) ||
146 test_bit(ICE_VF_STATE_DIS, vf->vf_states));
147 }
148
149 /**
150 * ice_wait_on_vf_reset - poll to make sure a given VF is ready after reset
151 * @vf: The VF being resseting
152 *
153 * The max poll time is about ~800ms, which is about the maximum time it takes
154 * for a VF to be reset and/or a VF driver to be removed.
155 */
ice_wait_on_vf_reset(struct ice_vf * vf)156 static void ice_wait_on_vf_reset(struct ice_vf *vf)
157 {
158 int i;
159
160 for (i = 0; i < ICE_MAX_VF_RESET_TRIES; i++) {
161 if (test_bit(ICE_VF_STATE_INIT, vf->vf_states))
162 break;
163 msleep(ICE_MAX_VF_RESET_SLEEP_MS);
164 }
165 }
166
167 /**
168 * ice_check_vf_ready_for_cfg - check if VF is ready to be configured/queried
169 * @vf: VF to check if it's ready to be configured/queried
170 *
171 * The purpose of this function is to make sure the VF is not in reset, not
172 * disabled, and initialized so it can be configured and/or queried by a host
173 * administrator.
174 */
ice_check_vf_ready_for_cfg(struct ice_vf * vf)175 int ice_check_vf_ready_for_cfg(struct ice_vf *vf)
176 {
177 ice_wait_on_vf_reset(vf);
178
179 if (ice_is_vf_disabled(vf))
180 return -EINVAL;
181
182 if (ice_check_vf_init(vf))
183 return -EBUSY;
184
185 return 0;
186 }
187
188 /**
189 * ice_trigger_vf_reset - Reset a VF on HW
190 * @vf: pointer to the VF structure
191 * @is_vflr: true if VFLR was issued, false if not
192 * @is_pfr: true if the reset was triggered due to a previous PFR
193 *
194 * Trigger hardware to start a reset for a particular VF. Expects the caller
195 * to wait the proper amount of time to allow hardware to reset the VF before
196 * it cleans up and restores VF functionality.
197 */
ice_trigger_vf_reset(struct ice_vf * vf,bool is_vflr,bool is_pfr)198 static void ice_trigger_vf_reset(struct ice_vf *vf, bool is_vflr, bool is_pfr)
199 {
200 /* Inform VF that it is no longer active, as a warning */
201 clear_bit(ICE_VF_STATE_ACTIVE, vf->vf_states);
202
203 /* Disable VF's configuration API during reset. The flag is re-enabled
204 * when it's safe again to access VF's VSI.
205 */
206 clear_bit(ICE_VF_STATE_INIT, vf->vf_states);
207
208 /* VF_MBX_ARQLEN and VF_MBX_ATQLEN are cleared by PFR, so the driver
209 * needs to clear them in the case of VFR/VFLR. If this is done for
210 * PFR, it can mess up VF resets because the VF driver may already
211 * have started cleanup by the time we get here.
212 */
213 if (!is_pfr)
214 vf->vf_ops->clear_mbx_register(vf);
215
216 vf->vf_ops->trigger_reset_register(vf, is_vflr);
217 }
218
ice_vf_clear_counters(struct ice_vf * vf)219 static void ice_vf_clear_counters(struct ice_vf *vf)
220 {
221 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
222
223 if (vsi)
224 vsi->num_vlan = 0;
225
226 vf->num_mac = 0;
227 memset(&vf->mdd_tx_events, 0, sizeof(vf->mdd_tx_events));
228 memset(&vf->mdd_rx_events, 0, sizeof(vf->mdd_rx_events));
229 }
230
231 /**
232 * ice_vf_pre_vsi_rebuild - tasks to be done prior to VSI rebuild
233 * @vf: VF to perform pre VSI rebuild tasks
234 *
235 * These tasks are items that don't need to be amortized since they are most
236 * likely called in a for loop with all VF(s) in the reset_all_vfs() case.
237 */
ice_vf_pre_vsi_rebuild(struct ice_vf * vf)238 static void ice_vf_pre_vsi_rebuild(struct ice_vf *vf)
239 {
240 /* Close any IRQ mapping now */
241 if (vf->vf_ops->irq_close)
242 vf->vf_ops->irq_close(vf);
243
244 ice_vf_clear_counters(vf);
245 vf->vf_ops->clear_reset_trigger(vf);
246 }
247
248 /**
249 * ice_vf_recreate_vsi - Release and re-create the VF's VSI
250 * @vf: VF to recreate the VSI for
251 *
252 * This is only called when a single VF is being reset (i.e. VVF, VFLR, host
253 * VF configuration change, etc)
254 *
255 * It releases and then re-creates a new VSI.
256 */
ice_vf_recreate_vsi(struct ice_vf * vf)257 static int ice_vf_recreate_vsi(struct ice_vf *vf)
258 {
259 struct ice_pf *pf = vf->pf;
260 int err;
261
262 ice_vf_vsi_release(vf);
263
264 err = vf->vf_ops->create_vsi(vf);
265 if (err) {
266 dev_err(ice_pf_to_dev(pf),
267 "Failed to recreate the VF%u's VSI, error %d\n",
268 vf->vf_id, err);
269 return err;
270 }
271
272 return 0;
273 }
274
275 /**
276 * ice_vf_rebuild_vsi - rebuild the VF's VSI
277 * @vf: VF to rebuild the VSI for
278 *
279 * This is only called when all VF(s) are being reset (i.e. PCIe Reset on the
280 * host, PFR, CORER, etc.).
281 *
282 * It reprograms the VSI configuration back into hardware.
283 */
ice_vf_rebuild_vsi(struct ice_vf * vf)284 static int ice_vf_rebuild_vsi(struct ice_vf *vf)
285 {
286 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
287 struct ice_pf *pf = vf->pf;
288
289 if (WARN_ON(!vsi))
290 return -EINVAL;
291
292 if (ice_vsi_rebuild(vsi, ICE_VSI_FLAG_INIT)) {
293 dev_err(ice_pf_to_dev(pf), "failed to rebuild VF %d VSI\n",
294 vf->vf_id);
295 return -EIO;
296 }
297 /* vsi->idx will remain the same in this case so don't update
298 * vf->lan_vsi_idx
299 */
300 vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx);
301 vf->lan_vsi_num = vsi->vsi_num;
302
303 return 0;
304 }
305
306 /**
307 * ice_vf_post_vsi_rebuild - Reset tasks that occur after VSI rebuild
308 * @vf: the VF being reset
309 *
310 * Perform reset tasks which must occur after the VSI has been re-created or
311 * rebuilt during a VF reset.
312 */
ice_vf_post_vsi_rebuild(struct ice_vf * vf)313 static void ice_vf_post_vsi_rebuild(struct ice_vf *vf)
314 {
315 ice_vf_rebuild_host_cfg(vf);
316 ice_vf_set_initialized(vf);
317
318 vf->vf_ops->post_vsi_rebuild(vf);
319 }
320
321 /**
322 * ice_is_any_vf_in_unicast_promisc - check if any VF(s)
323 * are in unicast promiscuous mode
324 * @pf: PF structure for accessing VF(s)
325 *
326 * Return false if no VF(s) are in unicast promiscuous mode,
327 * else return true
328 */
ice_is_any_vf_in_unicast_promisc(struct ice_pf * pf)329 bool ice_is_any_vf_in_unicast_promisc(struct ice_pf *pf)
330 {
331 bool is_vf_promisc = false;
332 struct ice_vf *vf;
333 unsigned int bkt;
334
335 rcu_read_lock();
336 ice_for_each_vf_rcu(pf, bkt, vf) {
337 /* found a VF that has promiscuous mode configured */
338 if (test_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states)) {
339 is_vf_promisc = true;
340 break;
341 }
342 }
343 rcu_read_unlock();
344
345 return is_vf_promisc;
346 }
347
348 /**
349 * ice_vf_get_promisc_masks - Calculate masks for promiscuous modes
350 * @vf: the VF pointer
351 * @vsi: the VSI to configure
352 * @ucast_m: promiscuous mask to apply to unicast
353 * @mcast_m: promiscuous mask to apply to multicast
354 *
355 * Decide which mask should be used for unicast and multicast filter,
356 * based on presence of VLANs
357 */
358 void
ice_vf_get_promisc_masks(struct ice_vf * vf,struct ice_vsi * vsi,u8 * ucast_m,u8 * mcast_m)359 ice_vf_get_promisc_masks(struct ice_vf *vf, struct ice_vsi *vsi,
360 u8 *ucast_m, u8 *mcast_m)
361 {
362 if (ice_vf_is_port_vlan_ena(vf) ||
363 ice_vsi_has_non_zero_vlans(vsi)) {
364 *mcast_m = ICE_MCAST_VLAN_PROMISC_BITS;
365 *ucast_m = ICE_UCAST_VLAN_PROMISC_BITS;
366 } else {
367 *mcast_m = ICE_MCAST_PROMISC_BITS;
368 *ucast_m = ICE_UCAST_PROMISC_BITS;
369 }
370 }
371
372 /**
373 * ice_vf_clear_all_promisc_modes - Clear promisc/allmulticast on VF VSI
374 * @vf: the VF pointer
375 * @vsi: the VSI to configure
376 *
377 * Clear all promiscuous/allmulticast filters for a VF
378 */
379 static int
ice_vf_clear_all_promisc_modes(struct ice_vf * vf,struct ice_vsi * vsi)380 ice_vf_clear_all_promisc_modes(struct ice_vf *vf, struct ice_vsi *vsi)
381 {
382 struct ice_pf *pf = vf->pf;
383 u8 ucast_m, mcast_m;
384 int ret = 0;
385
386 ice_vf_get_promisc_masks(vf, vsi, &ucast_m, &mcast_m);
387 if (test_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states)) {
388 if (!test_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, pf->flags)) {
389 if (ice_is_dflt_vsi_in_use(vsi->port_info))
390 ret = ice_clear_dflt_vsi(vsi);
391 } else {
392 ret = ice_vf_clear_vsi_promisc(vf, vsi, ucast_m);
393 }
394
395 if (ret) {
396 dev_err(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode failed\n");
397 } else {
398 clear_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states);
399 dev_info(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode succeeded\n");
400 }
401 }
402
403 if (test_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states)) {
404 ret = ice_vf_clear_vsi_promisc(vf, vsi, mcast_m);
405 if (ret) {
406 dev_err(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode failed\n");
407 } else {
408 clear_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states);
409 dev_info(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode succeeded\n");
410 }
411 }
412 return ret;
413 }
414
415 /**
416 * ice_vf_set_vsi_promisc - Enable promiscuous mode for a VF VSI
417 * @vf: the VF to configure
418 * @vsi: the VF's VSI
419 * @promisc_m: the promiscuous mode to enable
420 */
421 int
ice_vf_set_vsi_promisc(struct ice_vf * vf,struct ice_vsi * vsi,u8 promisc_m)422 ice_vf_set_vsi_promisc(struct ice_vf *vf, struct ice_vsi *vsi, u8 promisc_m)
423 {
424 struct ice_hw *hw = &vsi->back->hw;
425 int status;
426
427 if (ice_vf_is_port_vlan_ena(vf))
428 status = ice_fltr_set_vsi_promisc(hw, vsi->idx, promisc_m,
429 ice_vf_get_port_vlan_id(vf));
430 else if (ice_vsi_has_non_zero_vlans(vsi))
431 status = ice_fltr_set_vlan_vsi_promisc(hw, vsi, promisc_m);
432 else
433 status = ice_fltr_set_vsi_promisc(hw, vsi->idx, promisc_m, 0);
434
435 if (status && status != -EEXIST) {
436 dev_err(ice_pf_to_dev(vsi->back), "enable Tx/Rx filter promiscuous mode on VF-%u failed, error: %d\n",
437 vf->vf_id, status);
438 return status;
439 }
440
441 return 0;
442 }
443
444 /**
445 * ice_vf_clear_vsi_promisc - Disable promiscuous mode for a VF VSI
446 * @vf: the VF to configure
447 * @vsi: the VF's VSI
448 * @promisc_m: the promiscuous mode to disable
449 */
450 int
ice_vf_clear_vsi_promisc(struct ice_vf * vf,struct ice_vsi * vsi,u8 promisc_m)451 ice_vf_clear_vsi_promisc(struct ice_vf *vf, struct ice_vsi *vsi, u8 promisc_m)
452 {
453 struct ice_hw *hw = &vsi->back->hw;
454 int status;
455
456 if (ice_vf_is_port_vlan_ena(vf))
457 status = ice_fltr_clear_vsi_promisc(hw, vsi->idx, promisc_m,
458 ice_vf_get_port_vlan_id(vf));
459 else if (ice_vsi_has_non_zero_vlans(vsi))
460 status = ice_fltr_clear_vlan_vsi_promisc(hw, vsi, promisc_m);
461 else
462 status = ice_fltr_clear_vsi_promisc(hw, vsi->idx, promisc_m, 0);
463
464 if (status && status != -ENOENT) {
465 dev_err(ice_pf_to_dev(vsi->back), "disable Tx/Rx filter promiscuous mode on VF-%u failed, error: %d\n",
466 vf->vf_id, status);
467 return status;
468 }
469
470 return 0;
471 }
472
473 /**
474 * ice_reset_all_vfs - reset all allocated VFs in one go
475 * @pf: pointer to the PF structure
476 *
477 * Reset all VFs at once, in response to a PF or other device reset.
478 *
479 * First, tell the hardware to reset each VF, then do all the waiting in one
480 * chunk, and finally finish restoring each VF after the wait. This is useful
481 * during PF routines which need to reset all VFs, as otherwise it must perform
482 * these resets in a serialized fashion.
483 */
ice_reset_all_vfs(struct ice_pf * pf)484 void ice_reset_all_vfs(struct ice_pf *pf)
485 {
486 struct device *dev = ice_pf_to_dev(pf);
487 struct ice_hw *hw = &pf->hw;
488 struct ice_vf *vf;
489 unsigned int bkt;
490
491 /* If we don't have any VFs, then there is nothing to reset */
492 if (!ice_has_vfs(pf))
493 return;
494
495 mutex_lock(&pf->vfs.table_lock);
496
497 /* clear all malicious info if the VFs are getting reset */
498 ice_for_each_vf(pf, bkt, vf)
499 if (ice_mbx_clear_malvf(&hw->mbx_snapshot, pf->vfs.malvfs,
500 ICE_MAX_SRIOV_VFS, vf->vf_id))
501 dev_dbg(dev, "failed to clear malicious VF state for VF %u\n",
502 vf->vf_id);
503
504 /* If VFs have been disabled, there is no need to reset */
505 if (test_and_set_bit(ICE_VF_DIS, pf->state)) {
506 mutex_unlock(&pf->vfs.table_lock);
507 return;
508 }
509
510 /* Begin reset on all VFs at once */
511 ice_for_each_vf(pf, bkt, vf)
512 ice_trigger_vf_reset(vf, true, true);
513
514 /* HW requires some time to make sure it can flush the FIFO for a VF
515 * when it resets it. Now that we've triggered all of the VFs, iterate
516 * the table again and wait for each VF to complete.
517 */
518 ice_for_each_vf(pf, bkt, vf) {
519 if (!vf->vf_ops->poll_reset_status(vf)) {
520 /* Display a warning if at least one VF didn't manage
521 * to reset in time, but continue on with the
522 * operation.
523 */
524 dev_warn(dev, "VF %u reset check timeout\n", vf->vf_id);
525 break;
526 }
527 }
528
529 /* free VF resources to begin resetting the VSI state */
530 ice_for_each_vf(pf, bkt, vf) {
531 mutex_lock(&vf->cfg_lock);
532
533 vf->driver_caps = 0;
534 ice_vc_set_default_allowlist(vf);
535
536 ice_vf_fdir_exit(vf);
537 ice_vf_fdir_init(vf);
538 /* clean VF control VSI when resetting VFs since it should be
539 * setup only when VF creates its first FDIR rule.
540 */
541 if (vf->ctrl_vsi_idx != ICE_NO_VSI)
542 ice_vf_ctrl_invalidate_vsi(vf);
543
544 ice_vf_pre_vsi_rebuild(vf);
545 ice_vf_rebuild_vsi(vf);
546 ice_vf_post_vsi_rebuild(vf);
547
548 mutex_unlock(&vf->cfg_lock);
549 }
550
551 if (ice_is_eswitch_mode_switchdev(pf))
552 if (ice_eswitch_rebuild(pf))
553 dev_warn(dev, "eswitch rebuild failed\n");
554
555 ice_flush(hw);
556 clear_bit(ICE_VF_DIS, pf->state);
557
558 mutex_unlock(&pf->vfs.table_lock);
559 }
560
561 /**
562 * ice_notify_vf_reset - Notify VF of a reset event
563 * @vf: pointer to the VF structure
564 */
ice_notify_vf_reset(struct ice_vf * vf)565 static void ice_notify_vf_reset(struct ice_vf *vf)
566 {
567 struct ice_hw *hw = &vf->pf->hw;
568 struct virtchnl_pf_event pfe;
569
570 /* Bail out if VF is in disabled state, neither initialized, nor active
571 * state - otherwise proceed with notifications
572 */
573 if ((!test_bit(ICE_VF_STATE_INIT, vf->vf_states) &&
574 !test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states)) ||
575 test_bit(ICE_VF_STATE_DIS, vf->vf_states))
576 return;
577
578 pfe.event = VIRTCHNL_EVENT_RESET_IMPENDING;
579 pfe.severity = PF_EVENT_SEVERITY_CERTAIN_DOOM;
580 ice_aq_send_msg_to_vf(hw, vf->vf_id, VIRTCHNL_OP_EVENT,
581 VIRTCHNL_STATUS_SUCCESS, (u8 *)&pfe, sizeof(pfe),
582 NULL);
583 }
584
585 /**
586 * ice_reset_vf - Reset a particular VF
587 * @vf: pointer to the VF structure
588 * @flags: flags controlling behavior of the reset
589 *
590 * Flags:
591 * ICE_VF_RESET_VFLR - Indicates a reset is due to VFLR event
592 * ICE_VF_RESET_NOTIFY - Send VF a notification prior to reset
593 * ICE_VF_RESET_LOCK - Acquire VF cfg_lock before resetting
594 *
595 * Returns 0 if the VF is currently in reset, if resets are disabled, or if
596 * the VF resets successfully. Returns an error code if the VF fails to
597 * rebuild.
598 */
ice_reset_vf(struct ice_vf * vf,u32 flags)599 int ice_reset_vf(struct ice_vf *vf, u32 flags)
600 {
601 struct ice_pf *pf = vf->pf;
602 struct ice_vsi *vsi;
603 struct device *dev;
604 struct ice_hw *hw;
605 int err = 0;
606 bool rsd;
607
608 dev = ice_pf_to_dev(pf);
609 hw = &pf->hw;
610
611 if (flags & ICE_VF_RESET_NOTIFY)
612 ice_notify_vf_reset(vf);
613
614 if (test_bit(ICE_VF_RESETS_DISABLED, pf->state)) {
615 dev_dbg(dev, "Trying to reset VF %d, but all VF resets are disabled\n",
616 vf->vf_id);
617 return 0;
618 }
619
620 if (ice_is_vf_disabled(vf)) {
621 vsi = ice_get_vf_vsi(vf);
622 if (!vsi) {
623 dev_dbg(dev, "VF is already removed\n");
624 return -EINVAL;
625 }
626 ice_vsi_stop_lan_tx_rings(vsi, ICE_NO_RESET, vf->vf_id);
627
628 if (ice_vsi_is_rx_queue_active(vsi))
629 ice_vsi_stop_all_rx_rings(vsi);
630
631 dev_dbg(dev, "VF is already disabled, there is no need for resetting it, telling VM, all is fine %d\n",
632 vf->vf_id);
633 return 0;
634 }
635
636 if (flags & ICE_VF_RESET_LOCK)
637 mutex_lock(&vf->cfg_lock);
638 else
639 lockdep_assert_held(&vf->cfg_lock);
640
641 /* Set VF disable bit state here, before triggering reset */
642 set_bit(ICE_VF_STATE_DIS, vf->vf_states);
643 ice_trigger_vf_reset(vf, flags & ICE_VF_RESET_VFLR, false);
644
645 vsi = ice_get_vf_vsi(vf);
646 if (WARN_ON(!vsi)) {
647 err = -EIO;
648 goto out_unlock;
649 }
650
651 ice_dis_vf_qs(vf);
652
653 /* Call Disable LAN Tx queue AQ whether or not queues are
654 * enabled. This is needed for successful completion of VFR.
655 */
656 ice_dis_vsi_txq(vsi->port_info, vsi->idx, 0, 0, NULL, NULL,
657 NULL, vf->vf_ops->reset_type, vf->vf_id, NULL);
658
659 /* poll VPGEN_VFRSTAT reg to make sure
660 * that reset is complete
661 */
662 rsd = vf->vf_ops->poll_reset_status(vf);
663
664 /* Display a warning if VF didn't manage to reset in time, but need to
665 * continue on with the operation.
666 */
667 if (!rsd)
668 dev_warn(dev, "VF reset check timeout on VF %d\n", vf->vf_id);
669
670 vf->driver_caps = 0;
671 ice_vc_set_default_allowlist(vf);
672
673 /* disable promiscuous modes in case they were enabled
674 * ignore any error if disabling process failed
675 */
676 ice_vf_clear_all_promisc_modes(vf, vsi);
677
678 ice_eswitch_del_vf_mac_rule(vf);
679
680 ice_vf_fdir_exit(vf);
681 ice_vf_fdir_init(vf);
682 /* clean VF control VSI when resetting VF since it should be setup
683 * only when VF creates its first FDIR rule.
684 */
685 if (vf->ctrl_vsi_idx != ICE_NO_VSI)
686 ice_vf_ctrl_vsi_release(vf);
687
688 ice_vf_pre_vsi_rebuild(vf);
689
690 if (ice_vf_recreate_vsi(vf)) {
691 dev_err(dev, "Failed to release and setup the VF%u's VSI\n",
692 vf->vf_id);
693 err = -EFAULT;
694 goto out_unlock;
695 }
696
697 ice_vf_post_vsi_rebuild(vf);
698 vsi = ice_get_vf_vsi(vf);
699 if (WARN_ON(!vsi)) {
700 err = -EINVAL;
701 goto out_unlock;
702 }
703
704 ice_eswitch_update_repr(vsi);
705 ice_eswitch_replay_vf_mac_rule(vf);
706
707 /* if the VF has been reset allow it to come up again */
708 if (ice_mbx_clear_malvf(&hw->mbx_snapshot, pf->vfs.malvfs,
709 ICE_MAX_SRIOV_VFS, vf->vf_id))
710 dev_dbg(dev, "failed to clear malicious VF state for VF %u\n",
711 vf->vf_id);
712
713 out_unlock:
714 if (flags & ICE_VF_RESET_LOCK)
715 mutex_unlock(&vf->cfg_lock);
716
717 return err;
718 }
719
720 /**
721 * ice_set_vf_state_qs_dis - Set VF queues state to disabled
722 * @vf: pointer to the VF structure
723 */
ice_set_vf_state_qs_dis(struct ice_vf * vf)724 static void ice_set_vf_state_qs_dis(struct ice_vf *vf)
725 {
726 /* Clear Rx/Tx enabled queues flag */
727 bitmap_zero(vf->txq_ena, ICE_MAX_RSS_QS_PER_VF);
728 bitmap_zero(vf->rxq_ena, ICE_MAX_RSS_QS_PER_VF);
729 clear_bit(ICE_VF_STATE_QS_ENA, vf->vf_states);
730 }
731
732 /**
733 * ice_set_vf_state_dis - Set VF state to disabled
734 * @vf: pointer to the VF structure
735 */
ice_set_vf_state_dis(struct ice_vf * vf)736 void ice_set_vf_state_dis(struct ice_vf *vf)
737 {
738 ice_set_vf_state_qs_dis(vf);
739 vf->vf_ops->clear_reset_state(vf);
740 }
741
742 /* Private functions only accessed from other virtualization files */
743
744 /**
745 * ice_initialize_vf_entry - Initialize a VF entry
746 * @vf: pointer to the VF structure
747 */
ice_initialize_vf_entry(struct ice_vf * vf)748 void ice_initialize_vf_entry(struct ice_vf *vf)
749 {
750 struct ice_pf *pf = vf->pf;
751 struct ice_vfs *vfs;
752
753 vfs = &pf->vfs;
754
755 /* assign default capabilities */
756 vf->spoofchk = true;
757 vf->num_vf_qs = vfs->num_qps_per;
758 ice_vc_set_default_allowlist(vf);
759 ice_virtchnl_set_dflt_ops(vf);
760
761 /* ctrl_vsi_idx will be set to a valid value only when iAVF
762 * creates its first fdir rule.
763 */
764 ice_vf_ctrl_invalidate_vsi(vf);
765 ice_vf_fdir_init(vf);
766
767 mutex_init(&vf->cfg_lock);
768 }
769
770 /**
771 * ice_dis_vf_qs - Disable the VF queues
772 * @vf: pointer to the VF structure
773 */
ice_dis_vf_qs(struct ice_vf * vf)774 void ice_dis_vf_qs(struct ice_vf *vf)
775 {
776 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
777
778 if (WARN_ON(!vsi))
779 return;
780
781 ice_vsi_stop_lan_tx_rings(vsi, ICE_NO_RESET, vf->vf_id);
782 ice_vsi_stop_all_rx_rings(vsi);
783 ice_set_vf_state_qs_dis(vf);
784 }
785
786 /**
787 * ice_err_to_virt_err - translate errors for VF return code
788 * @err: error return code
789 */
ice_err_to_virt_err(int err)790 enum virtchnl_status_code ice_err_to_virt_err(int err)
791 {
792 switch (err) {
793 case 0:
794 return VIRTCHNL_STATUS_SUCCESS;
795 case -EINVAL:
796 case -ENODEV:
797 return VIRTCHNL_STATUS_ERR_PARAM;
798 case -ENOMEM:
799 return VIRTCHNL_STATUS_ERR_NO_MEMORY;
800 case -EALREADY:
801 case -EBUSY:
802 case -EIO:
803 case -ENOSPC:
804 return VIRTCHNL_STATUS_ERR_ADMIN_QUEUE_ERROR;
805 default:
806 return VIRTCHNL_STATUS_ERR_NOT_SUPPORTED;
807 }
808 }
809
810 /**
811 * ice_check_vf_init - helper to check if VF init complete
812 * @vf: the pointer to the VF to check
813 */
ice_check_vf_init(struct ice_vf * vf)814 int ice_check_vf_init(struct ice_vf *vf)
815 {
816 struct ice_pf *pf = vf->pf;
817
818 if (!test_bit(ICE_VF_STATE_INIT, vf->vf_states)) {
819 dev_err(ice_pf_to_dev(pf), "VF ID: %u in reset. Try again.\n",
820 vf->vf_id);
821 return -EBUSY;
822 }
823 return 0;
824 }
825
826 /**
827 * ice_vf_get_port_info - Get the VF's port info structure
828 * @vf: VF used to get the port info structure for
829 */
ice_vf_get_port_info(struct ice_vf * vf)830 struct ice_port_info *ice_vf_get_port_info(struct ice_vf *vf)
831 {
832 return vf->pf->hw.port_info;
833 }
834
835 /**
836 * ice_cfg_mac_antispoof - Configure MAC antispoof checking behavior
837 * @vsi: the VSI to configure
838 * @enable: whether to enable or disable the spoof checking
839 *
840 * Configure a VSI to enable (or disable) spoof checking behavior.
841 */
ice_cfg_mac_antispoof(struct ice_vsi * vsi,bool enable)842 static int ice_cfg_mac_antispoof(struct ice_vsi *vsi, bool enable)
843 {
844 struct ice_vsi_ctx *ctx;
845 int err;
846
847 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
848 if (!ctx)
849 return -ENOMEM;
850
851 ctx->info.sec_flags = vsi->info.sec_flags;
852 ctx->info.valid_sections = cpu_to_le16(ICE_AQ_VSI_PROP_SECURITY_VALID);
853
854 if (enable)
855 ctx->info.sec_flags |= ICE_AQ_VSI_SEC_FLAG_ENA_MAC_ANTI_SPOOF;
856 else
857 ctx->info.sec_flags &= ~ICE_AQ_VSI_SEC_FLAG_ENA_MAC_ANTI_SPOOF;
858
859 err = ice_update_vsi(&vsi->back->hw, vsi->idx, ctx, NULL);
860 if (err)
861 dev_err(ice_pf_to_dev(vsi->back), "Failed to configure Tx MAC anti-spoof %s for VSI %d, error %d\n",
862 enable ? "ON" : "OFF", vsi->vsi_num, err);
863 else
864 vsi->info.sec_flags = ctx->info.sec_flags;
865
866 kfree(ctx);
867
868 return err;
869 }
870
871 /**
872 * ice_vsi_ena_spoofchk - enable Tx spoof checking for this VSI
873 * @vsi: VSI to enable Tx spoof checking for
874 */
ice_vsi_ena_spoofchk(struct ice_vsi * vsi)875 static int ice_vsi_ena_spoofchk(struct ice_vsi *vsi)
876 {
877 struct ice_vsi_vlan_ops *vlan_ops;
878 int err = 0;
879
880 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi);
881
882 /* Allow VF with VLAN 0 only to send all tagged traffic */
883 if (vsi->type != ICE_VSI_VF || ice_vsi_has_non_zero_vlans(vsi)) {
884 err = vlan_ops->ena_tx_filtering(vsi);
885 if (err)
886 return err;
887 }
888
889 return ice_cfg_mac_antispoof(vsi, true);
890 }
891
892 /**
893 * ice_vsi_dis_spoofchk - disable Tx spoof checking for this VSI
894 * @vsi: VSI to disable Tx spoof checking for
895 */
ice_vsi_dis_spoofchk(struct ice_vsi * vsi)896 static int ice_vsi_dis_spoofchk(struct ice_vsi *vsi)
897 {
898 struct ice_vsi_vlan_ops *vlan_ops;
899 int err;
900
901 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi);
902
903 err = vlan_ops->dis_tx_filtering(vsi);
904 if (err)
905 return err;
906
907 return ice_cfg_mac_antispoof(vsi, false);
908 }
909
910 /**
911 * ice_vsi_apply_spoofchk - Apply Tx spoof checking setting to a VSI
912 * @vsi: VSI associated to the VF
913 * @enable: whether to enable or disable the spoof checking
914 */
ice_vsi_apply_spoofchk(struct ice_vsi * vsi,bool enable)915 int ice_vsi_apply_spoofchk(struct ice_vsi *vsi, bool enable)
916 {
917 int err;
918
919 if (enable)
920 err = ice_vsi_ena_spoofchk(vsi);
921 else
922 err = ice_vsi_dis_spoofchk(vsi);
923
924 return err;
925 }
926
927 /**
928 * ice_is_vf_trusted
929 * @vf: pointer to the VF info
930 */
ice_is_vf_trusted(struct ice_vf * vf)931 bool ice_is_vf_trusted(struct ice_vf *vf)
932 {
933 return test_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps);
934 }
935
936 /**
937 * ice_vf_has_no_qs_ena - check if the VF has any Rx or Tx queues enabled
938 * @vf: the VF to check
939 *
940 * Returns true if the VF has no Rx and no Tx queues enabled and returns false
941 * otherwise
942 */
ice_vf_has_no_qs_ena(struct ice_vf * vf)943 bool ice_vf_has_no_qs_ena(struct ice_vf *vf)
944 {
945 return (!bitmap_weight(vf->rxq_ena, ICE_MAX_RSS_QS_PER_VF) &&
946 !bitmap_weight(vf->txq_ena, ICE_MAX_RSS_QS_PER_VF));
947 }
948
949 /**
950 * ice_is_vf_link_up - check if the VF's link is up
951 * @vf: VF to check if link is up
952 */
ice_is_vf_link_up(struct ice_vf * vf)953 bool ice_is_vf_link_up(struct ice_vf *vf)
954 {
955 struct ice_port_info *pi = ice_vf_get_port_info(vf);
956
957 if (ice_check_vf_init(vf))
958 return false;
959
960 if (ice_vf_has_no_qs_ena(vf))
961 return false;
962 else if (vf->link_forced)
963 return vf->link_up;
964 else
965 return pi->phy.link_info.link_info &
966 ICE_AQ_LINK_UP;
967 }
968
969 /**
970 * ice_vf_set_host_trust_cfg - set trust setting based on pre-reset value
971 * @vf: VF to configure trust setting for
972 */
ice_vf_set_host_trust_cfg(struct ice_vf * vf)973 static void ice_vf_set_host_trust_cfg(struct ice_vf *vf)
974 {
975 if (vf->trusted)
976 set_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps);
977 else
978 clear_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps);
979 }
980
981 /**
982 * ice_vf_rebuild_host_mac_cfg - add broadcast and the VF's perm_addr/LAA
983 * @vf: VF to add MAC filters for
984 *
985 * Called after a VF VSI has been re-added/rebuilt during reset. The PF driver
986 * always re-adds a broadcast filter and the VF's perm_addr/LAA after reset.
987 */
ice_vf_rebuild_host_mac_cfg(struct ice_vf * vf)988 static int ice_vf_rebuild_host_mac_cfg(struct ice_vf *vf)
989 {
990 struct device *dev = ice_pf_to_dev(vf->pf);
991 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
992 u8 broadcast[ETH_ALEN];
993 int status;
994
995 if (WARN_ON(!vsi))
996 return -EINVAL;
997
998 if (ice_is_eswitch_mode_switchdev(vf->pf))
999 return 0;
1000
1001 eth_broadcast_addr(broadcast);
1002 status = ice_fltr_add_mac(vsi, broadcast, ICE_FWD_TO_VSI);
1003 if (status) {
1004 dev_err(dev, "failed to add broadcast MAC filter for VF %u, error %d\n",
1005 vf->vf_id, status);
1006 return status;
1007 }
1008
1009 vf->num_mac++;
1010
1011 if (is_valid_ether_addr(vf->hw_lan_addr)) {
1012 status = ice_fltr_add_mac(vsi, vf->hw_lan_addr,
1013 ICE_FWD_TO_VSI);
1014 if (status) {
1015 dev_err(dev, "failed to add default unicast MAC filter %pM for VF %u, error %d\n",
1016 &vf->hw_lan_addr[0], vf->vf_id,
1017 status);
1018 return status;
1019 }
1020 vf->num_mac++;
1021
1022 ether_addr_copy(vf->dev_lan_addr, vf->hw_lan_addr);
1023 }
1024
1025 return 0;
1026 }
1027
1028 /**
1029 * ice_vf_rebuild_host_vlan_cfg - add VLAN 0 filter or rebuild the Port VLAN
1030 * @vf: VF to add MAC filters for
1031 * @vsi: Pointer to VSI
1032 *
1033 * Called after a VF VSI has been re-added/rebuilt during reset. The PF driver
1034 * always re-adds either a VLAN 0 or port VLAN based filter after reset.
1035 */
ice_vf_rebuild_host_vlan_cfg(struct ice_vf * vf,struct ice_vsi * vsi)1036 static int ice_vf_rebuild_host_vlan_cfg(struct ice_vf *vf, struct ice_vsi *vsi)
1037 {
1038 struct ice_vsi_vlan_ops *vlan_ops = ice_get_compat_vsi_vlan_ops(vsi);
1039 struct device *dev = ice_pf_to_dev(vf->pf);
1040 int err;
1041
1042 if (ice_vf_is_port_vlan_ena(vf)) {
1043 err = vlan_ops->set_port_vlan(vsi, &vf->port_vlan_info);
1044 if (err) {
1045 dev_err(dev, "failed to configure port VLAN via VSI parameters for VF %u, error %d\n",
1046 vf->vf_id, err);
1047 return err;
1048 }
1049
1050 err = vlan_ops->add_vlan(vsi, &vf->port_vlan_info);
1051 } else {
1052 err = ice_vsi_add_vlan_zero(vsi);
1053 }
1054
1055 if (err) {
1056 dev_err(dev, "failed to add VLAN %u filter for VF %u during VF rebuild, error %d\n",
1057 ice_vf_is_port_vlan_ena(vf) ?
1058 ice_vf_get_port_vlan_id(vf) : 0, vf->vf_id, err);
1059 return err;
1060 }
1061
1062 err = vlan_ops->ena_rx_filtering(vsi);
1063 if (err)
1064 dev_warn(dev, "failed to enable Rx VLAN filtering for VF %d VSI %d during VF rebuild, error %d\n",
1065 vf->vf_id, vsi->idx, err);
1066
1067 return 0;
1068 }
1069
1070 /**
1071 * ice_vf_rebuild_host_tx_rate_cfg - re-apply the Tx rate limiting configuration
1072 * @vf: VF to re-apply the configuration for
1073 *
1074 * Called after a VF VSI has been re-added/rebuild during reset. The PF driver
1075 * needs to re-apply the host configured Tx rate limiting configuration.
1076 */
ice_vf_rebuild_host_tx_rate_cfg(struct ice_vf * vf)1077 static int ice_vf_rebuild_host_tx_rate_cfg(struct ice_vf *vf)
1078 {
1079 struct device *dev = ice_pf_to_dev(vf->pf);
1080 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
1081 int err;
1082
1083 if (WARN_ON(!vsi))
1084 return -EINVAL;
1085
1086 if (vf->min_tx_rate) {
1087 err = ice_set_min_bw_limit(vsi, (u64)vf->min_tx_rate * 1000);
1088 if (err) {
1089 dev_err(dev, "failed to set min Tx rate to %d Mbps for VF %u, error %d\n",
1090 vf->min_tx_rate, vf->vf_id, err);
1091 return err;
1092 }
1093 }
1094
1095 if (vf->max_tx_rate) {
1096 err = ice_set_max_bw_limit(vsi, (u64)vf->max_tx_rate * 1000);
1097 if (err) {
1098 dev_err(dev, "failed to set max Tx rate to %d Mbps for VF %u, error %d\n",
1099 vf->max_tx_rate, vf->vf_id, err);
1100 return err;
1101 }
1102 }
1103
1104 return 0;
1105 }
1106
1107 /**
1108 * ice_vf_rebuild_aggregator_node_cfg - rebuild aggregator node config
1109 * @vsi: Pointer to VSI
1110 *
1111 * This function moves VSI into corresponding scheduler aggregator node
1112 * based on cached value of "aggregator node info" per VSI
1113 */
ice_vf_rebuild_aggregator_node_cfg(struct ice_vsi * vsi)1114 static void ice_vf_rebuild_aggregator_node_cfg(struct ice_vsi *vsi)
1115 {
1116 struct ice_pf *pf = vsi->back;
1117 struct device *dev;
1118 int status;
1119
1120 if (!vsi->agg_node)
1121 return;
1122
1123 dev = ice_pf_to_dev(pf);
1124 if (vsi->agg_node->num_vsis == ICE_MAX_VSIS_IN_AGG_NODE) {
1125 dev_dbg(dev,
1126 "agg_id %u already has reached max_num_vsis %u\n",
1127 vsi->agg_node->agg_id, vsi->agg_node->num_vsis);
1128 return;
1129 }
1130
1131 status = ice_move_vsi_to_agg(pf->hw.port_info, vsi->agg_node->agg_id,
1132 vsi->idx, vsi->tc_cfg.ena_tc);
1133 if (status)
1134 dev_dbg(dev, "unable to move VSI idx %u into aggregator %u node",
1135 vsi->idx, vsi->agg_node->agg_id);
1136 else
1137 vsi->agg_node->num_vsis++;
1138 }
1139
1140 /**
1141 * ice_vf_rebuild_host_cfg - host admin configuration is persistent across reset
1142 * @vf: VF to rebuild host configuration on
1143 */
ice_vf_rebuild_host_cfg(struct ice_vf * vf)1144 void ice_vf_rebuild_host_cfg(struct ice_vf *vf)
1145 {
1146 struct device *dev = ice_pf_to_dev(vf->pf);
1147 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
1148
1149 if (WARN_ON(!vsi))
1150 return;
1151
1152 ice_vf_set_host_trust_cfg(vf);
1153
1154 if (ice_vf_rebuild_host_mac_cfg(vf))
1155 dev_err(dev, "failed to rebuild default MAC configuration for VF %d\n",
1156 vf->vf_id);
1157
1158 if (ice_vf_rebuild_host_vlan_cfg(vf, vsi))
1159 dev_err(dev, "failed to rebuild VLAN configuration for VF %u\n",
1160 vf->vf_id);
1161
1162 if (ice_vf_rebuild_host_tx_rate_cfg(vf))
1163 dev_err(dev, "failed to rebuild Tx rate limiting configuration for VF %u\n",
1164 vf->vf_id);
1165
1166 if (ice_vsi_apply_spoofchk(vsi, vf->spoofchk))
1167 dev_err(dev, "failed to rebuild spoofchk configuration for VF %d\n",
1168 vf->vf_id);
1169
1170 /* rebuild aggregator node config for main VF VSI */
1171 ice_vf_rebuild_aggregator_node_cfg(vsi);
1172 }
1173
1174 /**
1175 * ice_vf_ctrl_invalidate_vsi - invalidate ctrl_vsi_idx to remove VSI access
1176 * @vf: VF that control VSI is being invalidated on
1177 */
ice_vf_ctrl_invalidate_vsi(struct ice_vf * vf)1178 void ice_vf_ctrl_invalidate_vsi(struct ice_vf *vf)
1179 {
1180 vf->ctrl_vsi_idx = ICE_NO_VSI;
1181 }
1182
1183 /**
1184 * ice_vf_ctrl_vsi_release - invalidate the VF's control VSI after freeing it
1185 * @vf: VF that control VSI is being released on
1186 */
ice_vf_ctrl_vsi_release(struct ice_vf * vf)1187 void ice_vf_ctrl_vsi_release(struct ice_vf *vf)
1188 {
1189 ice_vsi_release(vf->pf->vsi[vf->ctrl_vsi_idx]);
1190 ice_vf_ctrl_invalidate_vsi(vf);
1191 }
1192
1193 /**
1194 * ice_vf_ctrl_vsi_setup - Set up a VF control VSI
1195 * @vf: VF to setup control VSI for
1196 *
1197 * Returns pointer to the successfully allocated VSI struct on success,
1198 * otherwise returns NULL on failure.
1199 */
ice_vf_ctrl_vsi_setup(struct ice_vf * vf)1200 struct ice_vsi *ice_vf_ctrl_vsi_setup(struct ice_vf *vf)
1201 {
1202 struct ice_vsi_cfg_params params = {};
1203 struct ice_pf *pf = vf->pf;
1204 struct ice_vsi *vsi;
1205
1206 params.type = ICE_VSI_CTRL;
1207 params.pi = ice_vf_get_port_info(vf);
1208 params.vf = vf;
1209 params.flags = ICE_VSI_FLAG_INIT;
1210
1211 vsi = ice_vsi_setup(pf, ¶ms);
1212 if (!vsi) {
1213 dev_err(ice_pf_to_dev(pf), "Failed to create VF control VSI\n");
1214 ice_vf_ctrl_invalidate_vsi(vf);
1215 }
1216
1217 return vsi;
1218 }
1219
1220 /**
1221 * ice_vf_init_host_cfg - Initialize host admin configuration
1222 * @vf: VF to initialize
1223 * @vsi: the VSI created at initialization
1224 *
1225 * Initialize the VF host configuration. Called during VF creation to setup
1226 * VLAN 0, add the VF VSI broadcast filter, and setup spoof checking. It
1227 * should only be called during VF creation.
1228 */
ice_vf_init_host_cfg(struct ice_vf * vf,struct ice_vsi * vsi)1229 int ice_vf_init_host_cfg(struct ice_vf *vf, struct ice_vsi *vsi)
1230 {
1231 struct ice_vsi_vlan_ops *vlan_ops;
1232 struct ice_pf *pf = vf->pf;
1233 u8 broadcast[ETH_ALEN];
1234 struct device *dev;
1235 int err;
1236
1237 dev = ice_pf_to_dev(pf);
1238
1239 err = ice_vsi_add_vlan_zero(vsi);
1240 if (err) {
1241 dev_warn(dev, "Failed to add VLAN 0 filter for VF %d\n",
1242 vf->vf_id);
1243 return err;
1244 }
1245
1246 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi);
1247 err = vlan_ops->ena_rx_filtering(vsi);
1248 if (err) {
1249 dev_warn(dev, "Failed to enable Rx VLAN filtering for VF %d\n",
1250 vf->vf_id);
1251 return err;
1252 }
1253
1254 eth_broadcast_addr(broadcast);
1255 err = ice_fltr_add_mac(vsi, broadcast, ICE_FWD_TO_VSI);
1256 if (err) {
1257 dev_err(dev, "Failed to add broadcast MAC filter for VF %d, status %d\n",
1258 vf->vf_id, err);
1259 return err;
1260 }
1261
1262 vf->num_mac = 1;
1263
1264 err = ice_vsi_apply_spoofchk(vsi, vf->spoofchk);
1265 if (err) {
1266 dev_warn(dev, "Failed to initialize spoofchk setting for VF %d\n",
1267 vf->vf_id);
1268 return err;
1269 }
1270
1271 return 0;
1272 }
1273
1274 /**
1275 * ice_vf_invalidate_vsi - invalidate vsi_idx/vsi_num to remove VSI access
1276 * @vf: VF to remove access to VSI for
1277 */
ice_vf_invalidate_vsi(struct ice_vf * vf)1278 void ice_vf_invalidate_vsi(struct ice_vf *vf)
1279 {
1280 vf->lan_vsi_idx = ICE_NO_VSI;
1281 vf->lan_vsi_num = ICE_NO_VSI;
1282 }
1283
1284 /**
1285 * ice_vf_vsi_release - Release the VF VSI and invalidate indexes
1286 * @vf: pointer to the VF structure
1287 *
1288 * Release the VF associated with this VSI and then invalidate the VSI
1289 * indexes.
1290 */
ice_vf_vsi_release(struct ice_vf * vf)1291 void ice_vf_vsi_release(struct ice_vf *vf)
1292 {
1293 struct ice_vsi *vsi = ice_get_vf_vsi(vf);
1294
1295 if (WARN_ON(!vsi))
1296 return;
1297
1298 ice_vsi_release(vsi);
1299 ice_vf_invalidate_vsi(vf);
1300 }
1301
1302 /**
1303 * ice_vf_set_initialized - VF is ready for VIRTCHNL communication
1304 * @vf: VF to set in initialized state
1305 *
1306 * After this function the VF will be ready to receive/handle the
1307 * VIRTCHNL_OP_GET_VF_RESOURCES message
1308 */
ice_vf_set_initialized(struct ice_vf * vf)1309 void ice_vf_set_initialized(struct ice_vf *vf)
1310 {
1311 ice_set_vf_state_qs_dis(vf);
1312 clear_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states);
1313 clear_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states);
1314 clear_bit(ICE_VF_STATE_DIS, vf->vf_states);
1315 set_bit(ICE_VF_STATE_INIT, vf->vf_states);
1316 memset(&vf->vlan_v2_caps, 0, sizeof(vf->vlan_v2_caps));
1317 }
1318