Home
last modified time | relevance | path

Searched refs:esw (Results 1 – 25 of 74) sorted by relevance

123

/linux/drivers/net/ethernet/mellanox/mlx5/core/
A Deswitch.c95 if (!esw || !MLX5_CAP_GEN(esw->dev, vport_group_manager)) in mlx5_eswitch_get_vport()
967 mlx5_eq_notifier_register(esw->dev, &esw->nb); in mlx5_eswitch_event_handlers_register()
981 mlx5_eq_notifier_unregister(esw->dev, &esw->nb); in mlx5_eswitch_event_handlers_unregister()
1281 esw->esw_funcs.num_vfs, esw->enabled_vports); in mlx5_eswitch_enable_locked()
1350 esw->esw_funcs.num_vfs, esw->enabled_vports); in mlx5_eswitch_disable_locked()
1545 esw = kzalloc(sizeof(*esw), GFP_KERNEL); in mlx5_eswitch_init()
1546 if (!esw) in mlx5_eswitch_init()
1604 kfree(esw); in mlx5_eswitch_init()
1610 if (!esw || !MLX5_VPORT_MANAGER(esw->dev)) in mlx5_eswitch_cleanup()
1627 kfree(esw); in mlx5_eswitch_cleanup()
[all …]
A Deswitch_offloads.c1850 esw_chains_destroy(esw, esw_chains(esw)); in esw_create_offloads_fdb_tables()
1879 esw_chains_destroy(esw, esw_chains(esw)); in esw_destroy_offloads_fdb_tables()
2320 err = esw->offloads.rep_ops[rep_type]->load(esw->dev, rep); in mlx5_esw_offloads_rep_load()
2536 vport = mlx5_eswitch_get_vport(esw, esw->manager_vport); in esw_set_master_egress_rule()
2807 esw); in esw_offloads_devcom_init()
3007 if (!esw || esw->mode != MLX5_ESWITCH_OFFLOADS) in mlx5_eswitch_reload_reps()
3107 mlx5_eswitch_unload_vf_vports(esw, esw->esw_funcs.num_vfs); in esw_vfs_changed_event_handler()
3126 esw = host_work->esw; in esw_functions_changed_event_handler()
3151 host_work->esw = esw; in mlx5_esw_funcs_changed_handler()
3390 if (IS_ERR(esw)) in mlx5_devlink_eswitch_mode_set()
[all …]
A Deswitch.h84 #define esw_chains(esw) \ argument
85 ((esw)->fdb_table.offloads.esw_chains_priv)
271 struct mlx5_eswitch *esw; member
331 int esw_offloads_enable(struct mlx5_eswitch *esw);
521 return esw->qos.enabled; in mlx5_esw_qos_enabled()
557 return esw && MLX5_ESWITCH_MANAGER(esw->dev); in mlx5_esw_allowed()
570 return esw->manager_vport == vport_num; in mlx5_esw_is_manager_vport()
615 xa_for_each(&((esw)->vports), index, vport)
723 int mlx5_esw_try_lock(struct mlx5_eswitch *esw);
724 void mlx5_esw_unlock(struct mlx5_eswitch *esw);
[all …]
A Deswitch_offloads_termtbl.c120 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create()
122 hash_for_each_possible(esw->offloads.termtbl_tbl, tt, in mlx5_eswitch_termtbl_get_create()
152 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create()
156 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create()
161 mlx5_eswitch_termtbl_put(struct mlx5_eswitch *esw, in mlx5_eswitch_termtbl_put() argument
164 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put()
167 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put()
200 if (MLX5_CAP_ESW_FLOWTABLE(esw->dev, flow_source)) in mlx5_eswitch_offload_is_uplink_port()
212 mlx5_eswitch_termtbl_required(struct mlx5_eswitch *esw, in mlx5_eswitch_termtbl_required() argument
240 mlx5_eswitch_add_termtbl_rule(struct mlx5_eswitch *esw, in mlx5_eswitch_add_termtbl_rule() argument
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/
A Dlegacy.c53 if (!esw->fdb_table.legacy.fdb) in esw_destroy_legacy_fdb_table()
208 esw_cleanup_vepa_rules(esw); in esw_destroy_legacy_table()
227 mlx5_esw_for_each_vf_vport(esw, i, vport, esw->esw_funcs.num_vfs) in esw_legacy_enable()
316 if (!esw) in mlx5_eswitch_set_vepa()
319 if (!mlx5_esw_allowed(esw)) in mlx5_eswitch_set_vepa()
337 if (!esw) in mlx5_eswitch_get_vepa()
340 if (!mlx5_esw_allowed(esw)) in mlx5_eswitch_get_vepa()
433 if (!mlx5_esw_allowed(esw)) in mlx5_eswitch_set_vport_vlan()
462 if (!mlx5_esw_allowed(esw)) in mlx5_eswitch_set_vport_spoofchk()
475 mlx5_core_warn(esw->dev, in mlx5_eswitch_set_vport_spoofchk()
[all …]
A Dqos.c89 esw_warn(esw->dev, in esw_qos_vport_config()
543 if (esw->qos.enabled) in mlx5_esw_qos_create()
563 esw->qos.group0 = esw_qos_create_rate_group(esw, NULL); in mlx5_esw_qos_create()
591 if (!esw->qos.enabled) in mlx5_esw_qos_destroy()
594 if (esw->qos.group0) in mlx5_esw_qos_destroy()
595 esw_qos_destroy_rate_group(esw, esw->qos.group0, NULL); in mlx5_esw_qos_destroy()
614 if (!esw->qos.enabled) in mlx5_esw_qos_vport_enable()
801 if (IS_ERR(esw)) in mlx5_esw_devlink_rate_node_new()
802 return PTR_ERR(esw); in mlx5_esw_devlink_rate_node_new()
832 if (IS_ERR(esw)) in mlx5_esw_devlink_rate_node_del()
[all …]
A Dvporttbl.c54 key->vhca_id = MLX5_CAP_GEN(esw->dev, vhca_id); in flow_attr_to_vport_key()
75 struct mlx5_core_dev *dev = esw->dev; in mlx5_esw_vporttbl_get()
82 mutex_lock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get()
83 hkey = flow_attr_to_vport_key(esw, attr, &skey); in mlx5_esw_vporttbl_get()
84 e = esw_vport_tbl_lookup(esw, &skey, hkey); in mlx5_esw_vporttbl_get()
112 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get()
118 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get()
129 mutex_lock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_put()
130 hkey = flow_attr_to_vport_key(esw, attr, &key); in mlx5_esw_vporttbl_put()
131 e = esw_vport_tbl_lookup(esw, &key, hkey); in mlx5_esw_vporttbl_put()
[all …]
A Dbridge.c171 esw_warn(esw->dev, in mlx5_esw_bridge_ingress_vlan_fg_create()
351 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_ingress_table_init() local
359 esw); in mlx5_esw_bridge_ingress_table_init()
365 esw); in mlx5_esw_bridge_ingress_table_init()
434 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_egress_table_init() local
440 esw); in mlx5_esw_bridge_egress_table_init()
1071 struct mlx5_eswitch *esw = bridge->br_offloads->esw; in mlx5_esw_bridge_vlan_flush() local
1158 esw); in mlx5_esw_bridge_fdb_entry_init()
1294 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_vport_init() local
1518 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_fdb_remove() local
[all …]
A Ddevlink_port.c21 mlx5_eswitch_is_vf_vport(esw, vport_num); in mlx5_esw_devlink_port_supported()
26 struct mlx5_core_dev *dev = esw->dev; in mlx5_esw_dl_port_alloc()
70 struct mlx5_core_dev *dev = esw->dev; in mlx5_esw_offloads_devlink_port_register()
77 if (!mlx5_esw_devlink_port_supported(esw, vport_num)) in mlx5_esw_offloads_devlink_port_register()
80 vport = mlx5_eswitch_get_vport(esw, vport_num); in mlx5_esw_offloads_devlink_port_register()
84 dl_port = mlx5_esw_dl_port_alloc(esw, vport_num); in mlx5_esw_offloads_devlink_port_register()
115 vport = mlx5_eswitch_get_vport(esw, vport_num); in mlx5_esw_offloads_devlink_port_unregister()
133 vport = mlx5_eswitch_get_vport(esw, vport_num); in mlx5_esw_offloads_devlink_port()
140 struct mlx5_core_dev *dev = esw->dev; in mlx5_esw_devlink_sf_port_register()
148 vport = mlx5_eswitch_get_vport(esw, vport_num); in mlx5_esw_devlink_sf_port_register()
[all …]
A Dindir_table.c74 mlx5_esw_indir_table_needed(struct mlx5_eswitch *esw, in mlx5_esw_indir_table_needed() argument
85 mlx5_eswitch_is_vf_vport(esw, vport_num) && in mlx5_esw_indir_table_needed()
86 esw->dev == dest_mdev && in mlx5_esw_indir_table_needed()
123 struct mlx5_fs_chains *chains = esw_chains(esw); in mlx5_esw_indir_table_rule_get()
260 struct mlx5_fs_chains *chains = esw_chains(esw); in mlx5_esw_indir_table_rule_put()
277 mlx5_modify_header_dealloc(esw->dev, rule->mh); in mlx5_esw_indir_table_rule_put()
423 err = mlx5_create_indir_fwd_group(esw, e); in mlx5_esw_indir_table_entry_create()
434 mlx5_esw_indir_table_rule_put(esw, attr, e); in mlx5_esw_indir_table_entry_create()
465 mutex_lock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_get()
497 mutex_lock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_put()
[all …]
A Dindir_table.h14 struct mlx5_flow_table *mlx5_esw_indir_table_get(struct mlx5_eswitch *esw,
18 void mlx5_esw_indir_table_put(struct mlx5_eswitch *esw,
23 mlx5_esw_indir_table_needed(struct mlx5_eswitch *esw,
45 mlx5_esw_indir_table_get(struct mlx5_eswitch *esw, in mlx5_esw_indir_table_get() argument
54 mlx5_esw_indir_table_put(struct mlx5_eswitch *esw, in mlx5_esw_indir_table_put() argument
61 mlx5_esw_indir_table_needed(struct mlx5_eswitch *esw, in mlx5_esw_indir_table_needed() argument
A Dqos.h9 int mlx5_esw_qos_set_vport_min_rate(struct mlx5_eswitch *esw,
13 int mlx5_esw_qos_set_vport_max_rate(struct mlx5_eswitch *esw,
17 void mlx5_esw_qos_create(struct mlx5_eswitch *esw);
18 void mlx5_esw_qos_destroy(struct mlx5_eswitch *esw);
19 int mlx5_esw_qos_vport_enable(struct mlx5_eswitch *esw, struct mlx5_vport *vport,
21 void mlx5_esw_qos_vport_disable(struct mlx5_eswitch *esw, struct mlx5_vport *vport);
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/acl/
A Dingress_ofld.c10 esw_acl_ingress_prio_tag_enabled(struct mlx5_eswitch *esw, in esw_acl_ingress_prio_tag_enabled() argument
13 return (MLX5_CAP_GEN(esw->dev, prio_tag_required) && in esw_acl_ingress_prio_tag_enabled()
14 mlx5_eswitch_is_vf_vport(esw, vport->vport)); in esw_acl_ingress_prio_tag_enabled()
52 esw_warn(esw->dev, in esw_acl_ingress_prio_tag_create()
87 esw_warn(esw->dev, in esw_acl_ingress_mod_metadata_create()
100 esw_warn(esw->dev, in esw_acl_ingress_mod_metadata_create()
128 esw_warn(esw->dev, in esw_acl_ingress_ofld_rules_create()
138 esw_warn(esw->dev, in esw_acl_ingress_ofld_rules_create()
271 esw_debug(esw->dev, in esw_acl_ingress_ofld_setup()
290 esw_acl_ingress_ofld_rules_destroy(esw, vport); in esw_acl_ingress_ofld_cleanup()
[all …]
A Degress_ofld.c47 esw_warn(esw->dev, in esw_acl_egress_ofld_fwd2vport_create()
63 if (MLX5_CAP_GEN(esw->dev, prio_tag_required)) { in esw_acl_egress_ofld_rules_create()
68 esw_debug(esw->dev, in esw_acl_egress_ofld_rules_create()
111 if (MLX5_CAP_GEN(esw->dev, prio_tag_required)) { in esw_acl_egress_ofld_groups_create()
112 ret = esw_acl_egress_vlan_grp_create(esw, vport); in esw_acl_egress_ofld_groups_create()
119 if (!mlx5_esw_acl_egress_fwd2vport_supported(esw)) in esw_acl_egress_ofld_groups_create()
136 esw_warn(esw->dev, in esw_acl_egress_ofld_groups_create()
169 return mlx5_eswitch_is_vf_vport(esw, vport_num) || mlx5_esw_is_sf_vport(esw, vport_num); in esw_acl_egress_needed()
178 !MLX5_CAP_GEN(esw->dev, prio_tag_required)) in esw_acl_egress_ofld_setup()
181 if (!esw_acl_egress_needed(esw, vport->vport)) in esw_acl_egress_ofld_setup()
[all …]
A Dofld.h10 int esw_acl_egress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport);
12 int mlx5_esw_acl_egress_vport_bond(struct mlx5_eswitch *esw, u16 active_vport_num,
14 int mlx5_esw_acl_egress_vport_unbond(struct mlx5_eswitch *esw, u16 vport_num);
16 static inline bool mlx5_esw_acl_egress_fwd2vport_supported(struct mlx5_eswitch *esw) in mlx5_esw_acl_egress_fwd2vport_supported() argument
18 return esw && esw->mode == MLX5_ESWITCH_OFFLOADS && in mlx5_esw_acl_egress_fwd2vport_supported()
19 mlx5_eswitch_vport_match_metadata_enabled(esw) && in mlx5_esw_acl_egress_fwd2vport_supported()
20 MLX5_CAP_ESW_FLOWTABLE(esw->dev, egress_acl_forward_to_vport); in mlx5_esw_acl_egress_fwd2vport_supported()
24 int esw_acl_ingress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport);
25 void esw_acl_ingress_ofld_cleanup(struct mlx5_eswitch *esw, struct mlx5_vport *vport);
26 int mlx5_esw_acl_ingress_vport_bond_update(struct mlx5_eswitch *esw, u16 vport_num,
A Degress_lgcy.c22 struct mlx5_core_dev *dev = esw->dev; in esw_acl_egress_lgcy_groups_create()
27 err = esw_acl_egress_vlan_grp_create(esw, vport); in esw_acl_egress_lgcy_groups_create()
67 int esw_acl_egress_lgcy_setup(struct mlx5_eswitch *esw, in esw_acl_egress_lgcy_setup() argument
85 drop_counter = mlx5_fc_create(esw->dev, false); in esw_acl_egress_lgcy_setup()
87 esw_warn(esw->dev, in esw_acl_egress_lgcy_setup()
98 esw_acl_egress_lgcy_cleanup(esw, vport); in esw_acl_egress_lgcy_setup()
103 vport->egress.acl = esw_acl_table_create(esw, vport, in esw_acl_egress_lgcy_setup()
112 err = esw_acl_egress_lgcy_groups_create(esw, vport); in esw_acl_egress_lgcy_setup()
117 esw_debug(esw->dev, in esw_acl_egress_lgcy_setup()
142 esw_warn(esw->dev, in esw_acl_egress_lgcy_setup()
[all …]
A Dingress_lgcy.c22 struct mlx5_core_dev *dev = esw->dev; in esw_acl_ingress_lgcy_groups_create()
139 int esw_acl_ingress_lgcy_setup(struct mlx5_eswitch *esw, in esw_acl_ingress_lgcy_setup() argument
166 counter = mlx5_fc_create(esw->dev, false); in esw_acl_ingress_lgcy_setup()
168 esw_warn(esw->dev, in esw_acl_ingress_lgcy_setup()
177 esw_acl_ingress_lgcy_cleanup(esw, vport); in esw_acl_ingress_lgcy_setup()
182 vport->ingress.acl = esw_acl_table_create(esw, vport, in esw_acl_ingress_lgcy_setup()
191 err = esw_acl_ingress_lgcy_groups_create(esw, vport); in esw_acl_ingress_lgcy_setup()
196 esw_debug(esw->dev, in esw_acl_ingress_lgcy_setup()
228 esw_warn(esw->dev, in esw_acl_ingress_lgcy_setup()
250 esw_warn(esw->dev, in esw_acl_ingress_lgcy_setup()
[all …]
/linux/drivers/s390/cio/
A Ddevice_status.c124 cdev_irb->esw.esw1.lpum = irb->esw.esw1.lpum; in ccw_device_accumulate_esw()
158 cdev_irb->esw.esw0.erw.cpt = irb->esw.esw0.erw.cpt; in ccw_device_accumulate_esw()
160 cdev_irb->esw.esw0.erw.fsavf = irb->esw.esw0.erw.fsavf; in ccw_device_accumulate_esw()
163 memcpy(cdev_irb->esw.esw0.faddr, irb->esw.esw0.faddr, in ccw_device_accumulate_esw()
166 cdev_irb->esw.esw0.erw.fsaf = irb->esw.esw0.erw.fsaf; in ccw_device_accumulate_esw()
169 cdev_irb->esw.esw0.erw.scavf = irb->esw.esw0.erw.scavf; in ccw_device_accumulate_esw()
172 cdev_irb->esw.esw0.saddr = irb->esw.esw0.saddr; in ccw_device_accumulate_esw()
178 cdev_irb->esw.esw0.erw.auth = irb->esw.esw0.erw.auth; in ccw_device_accumulate_esw()
180 cdev_irb->esw.esw0.erw.pvrf = irb->esw.esw0.erw.pvrf; in ccw_device_accumulate_esw()
184 cdev_irb->esw.esw0.erw.cons = irb->esw.esw0.erw.cons; in ccw_device_accumulate_esw()
[all …]
/linux/include/linux/mlx5/
A Deswitch.h42 int (*event)(struct mlx5_eswitch *esw,
60 struct mlx5_eswitch *esw; member
63 void mlx5_eswitch_register_vport_reps(struct mlx5_eswitch *esw,
67 void *mlx5_eswitch_get_proto_dev(struct mlx5_eswitch *esw,
70 struct mlx5_eswitch_rep *mlx5_eswitch_vport_rep(struct mlx5_eswitch *esw,
82 bool mlx5_eswitch_reg_c1_loopback_enabled(const struct mlx5_eswitch *esw);
105 u32 mlx5_eswitch_get_vport_metadata_for_match(struct mlx5_eswitch *esw,
107 u32 mlx5_eswitch_get_vport_metadata_for_set(struct mlx5_eswitch *esw,
150 struct mlx5_core_dev *mlx5_eswitch_get_core_dev(struct mlx5_eswitch *esw);
166 mlx5_eswitch_reg_c1_loopback_enabled(const struct mlx5_eswitch *esw) in mlx5_eswitch_reg_c1_loopback_enabled() argument
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/rep/
A Dbridge.c26 return esw == priv->mdev->priv.eswitch; in mlx5_esw_bridge_dev_same_esw()
36 esw_mdev = esw->dev; in mlx5_esw_bridge_dev_same_hw()
113 struct mlx5_eswitch *esw) in mlx5_esw_bridge_is_local() argument
136 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_port_changeupper() local
381 struct mlx5_eswitch *esw = br_offloads->esw; in mlx5_esw_bridge_switchdev_event() local
461 struct mlx5_eswitch *esw = in mlx5e_rep_bridge_init() local
466 br_offloads = mlx5_esw_bridge_init(esw); in mlx5e_rep_bridge_init()
512 mlx5_esw_bridge_cleanup(esw); in mlx5e_rep_bridge_init()
519 struct mlx5_eswitch *esw = in mlx5e_rep_bridge_cleanup() local
522 br_offloads = esw->br_offloads; in mlx5e_rep_bridge_cleanup()
[all …]
A Dtc.c95 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_update_flows() local
101 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_rep_update_flows()
127 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_rep_update_flows()
191 struct mlx5_eswitch *esw; in mlx5e_rep_setup_ft_cb() local
198 esw = priv->mdev->priv.eswitch; in mlx5e_rep_setup_ft_cb()
373 struct mlx5_eswitch *esw; in mlx5e_rep_indr_setup_ft_cb() local
378 esw = mpriv->mdev->priv.eswitch; in mlx5e_rep_indr_setup_ft_cb()
653 struct mlx5_eswitch *esw; in mlx5e_restore_skb_chain() local
663 esw = priv->mdev->priv.eswitch; in mlx5e_restore_skb_chain()
729 struct mlx5_eswitch *esw; in mlx5e_rep_tc_receive() local
[all …]
/linux/drivers/infiniband/hw/mlx5/
A Dib_rep.c25 mlx5_ib_get_rep_netdev(rep->esw, rep->vport); in mlx5_ib_set_vport_rep()
139 struct mlx5_eswitch *esw; in mlx5_ib_vport_rep_unload() local
143 esw = peer_mdev->priv.eswitch; in mlx5_ib_vport_rep_unload()
159 struct mlx5_eswitch *esw; in mlx5_ib_register_peer_vport_reps() local
164 esw = peer_mdev->priv.eswitch; in mlx5_ib_register_peer_vport_reps()
189 return mlx5_eswitch_add_send_to_vport_rule(esw, esw, rep, sq->base.mqp.qpn); in create_flow_rule_vport_sq()
197 struct mlx5_eswitch *esw; in mlx5r_rep_probe() local
199 esw = mdev->priv.eswitch; in mlx5r_rep_probe()
208 struct mlx5_eswitch *esw; in mlx5r_rep_remove() local
210 esw = mdev->priv.eswitch; in mlx5r_rep_remove()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
A Dtc_tun_encap.c386 struct mlx5_eswitch *esw; in mlx5e_tc_update_neigh_used_value() local
389 esw = priv->mdev->priv.eswitch; in mlx5e_tc_update_neigh_used_value()
390 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_tc_update_neigh_used_value()
504 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_detach_encap()
527 mutex_lock(&esw->offloads.decap_tbl_lock); in mlx5e_detach_decap()
773 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_attach_encap()
828 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_attach_encap()
903 mutex_lock(&esw->offloads.decap_tbl_lock); in mlx5e_attach_decap()
940 mutex_lock(&esw->offloads.decap_tbl_lock); in mlx5e_attach_decap()
1532 struct mlx5_eswitch *esw; in mlx5e_tc_fib_event_work() local
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/tc/
A Dsample.c22 struct mlx5_eswitch *esw; member
63 struct mlx5_eswitch *esw = tc_psample->esw; in sampler_termtbl_create() local
66 struct mlx5_core_dev *dev = esw->dev; in sampler_termtbl_create()
95 dest.vport.num = esw->manager_vport; in sampler_termtbl_create()
285 struct mlx5_eswitch *esw = tc_psample->esw; in sample_restore_get() local
286 struct mlx5_core_dev *mdev = esw->dev; in sample_restore_get()
504 struct mlx5_eswitch *esw; in mlx5e_tc_sample_offload() local
523 esw = tc_psample->esw; in mlx5e_tc_sample_offload()
626 struct mlx5_eswitch *esw; in mlx5e_tc_sample_unoffload() local
634 esw = tc_psample->esw; in mlx5e_tc_sample_unoffload()
[all …]
A Dint_port.c33 return mlx5_eswitch_vport_match_metadata_enabled(esw) && in mlx5e_tc_int_port_supported()
34 MLX5_CAP_GEN(esw->dev, reg_c_preserve); in mlx5e_tc_int_port_supported()
61 mlx5e_int_port_create_rx_rule(struct mlx5_eswitch *esw, in mlx5e_int_port_create_rx_rule() argument
154 struct mlx5_eswitch *esw = priv->dev->priv.eswitch; in mlx5e_int_port_add() local
182 ctx = esw->offloads.reg_c0_obj_pool; in mlx5e_int_port_add()
230 struct mlx5_eswitch *esw = priv->dev->priv.eswitch; in mlx5e_int_port_remove() local
233 ctx = esw->offloads.reg_c0_obj_pool; in mlx5e_int_port_remove()
308 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_int_port_init() local
312 if (!mlx5e_tc_int_port_supported(esw)) in mlx5e_tc_int_port_init()
360 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_int_port_init_rep_rx() local
[all …]

Completed in 78 milliseconds

123