1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/kernel.h>
3 #include <linux/list.h>
4 #include <linux/netdevice.h>
5 #include <linux/rtnetlink.h>
6 #include <linux/skbuff.h>
7 #include <net/ip.h>
8 #include <net/switchdev.h>
9
10 #include "br_private.h"
11
12 static struct static_key_false br_switchdev_tx_fwd_offload;
13
nbp_switchdev_can_offload_tx_fwd(const struct net_bridge_port * p,const struct sk_buff * skb)14 static bool nbp_switchdev_can_offload_tx_fwd(const struct net_bridge_port *p,
15 const struct sk_buff *skb)
16 {
17 if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
18 return false;
19
20 return (p->flags & BR_TX_FWD_OFFLOAD) &&
21 (p->hwdom != BR_INPUT_SKB_CB(skb)->src_hwdom);
22 }
23
br_switchdev_frame_uses_tx_fwd_offload(struct sk_buff * skb)24 bool br_switchdev_frame_uses_tx_fwd_offload(struct sk_buff *skb)
25 {
26 if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
27 return false;
28
29 return BR_INPUT_SKB_CB(skb)->tx_fwd_offload;
30 }
31
br_switchdev_frame_set_offload_fwd_mark(struct sk_buff * skb)32 void br_switchdev_frame_set_offload_fwd_mark(struct sk_buff *skb)
33 {
34 skb->offload_fwd_mark = br_switchdev_frame_uses_tx_fwd_offload(skb);
35 }
36
37 /* Mark the frame for TX forwarding offload if this egress port supports it */
nbp_switchdev_frame_mark_tx_fwd_offload(const struct net_bridge_port * p,struct sk_buff * skb)38 void nbp_switchdev_frame_mark_tx_fwd_offload(const struct net_bridge_port *p,
39 struct sk_buff *skb)
40 {
41 if (nbp_switchdev_can_offload_tx_fwd(p, skb))
42 BR_INPUT_SKB_CB(skb)->tx_fwd_offload = true;
43 }
44
45 /* Lazily adds the hwdom of the egress bridge port to the bit mask of hwdoms
46 * that the skb has been already forwarded to, to avoid further cloning to
47 * other ports in the same hwdom by making nbp_switchdev_allowed_egress()
48 * return false.
49 */
nbp_switchdev_frame_mark_tx_fwd_to_hwdom(const struct net_bridge_port * p,struct sk_buff * skb)50 void nbp_switchdev_frame_mark_tx_fwd_to_hwdom(const struct net_bridge_port *p,
51 struct sk_buff *skb)
52 {
53 if (nbp_switchdev_can_offload_tx_fwd(p, skb))
54 set_bit(p->hwdom, &BR_INPUT_SKB_CB(skb)->fwd_hwdoms);
55 }
56
nbp_switchdev_frame_mark(const struct net_bridge_port * p,struct sk_buff * skb)57 void nbp_switchdev_frame_mark(const struct net_bridge_port *p,
58 struct sk_buff *skb)
59 {
60 if (p->hwdom)
61 BR_INPUT_SKB_CB(skb)->src_hwdom = p->hwdom;
62 }
63
nbp_switchdev_allowed_egress(const struct net_bridge_port * p,const struct sk_buff * skb)64 bool nbp_switchdev_allowed_egress(const struct net_bridge_port *p,
65 const struct sk_buff *skb)
66 {
67 struct br_input_skb_cb *cb = BR_INPUT_SKB_CB(skb);
68
69 return !test_bit(p->hwdom, &cb->fwd_hwdoms) &&
70 (!skb->offload_fwd_mark || cb->src_hwdom != p->hwdom);
71 }
72
73 /* Flags that can be offloaded to hardware */
74 #define BR_PORT_FLAGS_HW_OFFLOAD (BR_LEARNING | BR_FLOOD | BR_PORT_MAB | \
75 BR_MCAST_FLOOD | BR_BCAST_FLOOD | BR_PORT_LOCKED | \
76 BR_HAIRPIN_MODE | BR_ISOLATED | BR_MULTICAST_TO_UNICAST)
77
br_switchdev_set_port_flag(struct net_bridge_port * p,unsigned long flags,unsigned long mask,struct netlink_ext_ack * extack)78 int br_switchdev_set_port_flag(struct net_bridge_port *p,
79 unsigned long flags,
80 unsigned long mask,
81 struct netlink_ext_ack *extack)
82 {
83 struct switchdev_attr attr = {
84 .orig_dev = p->dev,
85 };
86 struct switchdev_notifier_port_attr_info info = {
87 .attr = &attr,
88 };
89 int err;
90
91 mask &= BR_PORT_FLAGS_HW_OFFLOAD;
92 if (!mask)
93 return 0;
94
95 attr.id = SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS;
96 attr.u.brport_flags.val = flags;
97 attr.u.brport_flags.mask = mask;
98
99 /* We run from atomic context here */
100 err = call_switchdev_notifiers(SWITCHDEV_PORT_ATTR_SET, p->dev,
101 &info.info, extack);
102 err = notifier_to_errno(err);
103 if (err == -EOPNOTSUPP)
104 return 0;
105
106 if (err) {
107 NL_SET_ERR_MSG_WEAK_MOD(extack,
108 "bridge flag offload is not supported");
109 return -EOPNOTSUPP;
110 }
111
112 attr.id = SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS;
113 attr.flags = SWITCHDEV_F_DEFER;
114
115 err = switchdev_port_attr_set(p->dev, &attr, extack);
116 if (err) {
117 NL_SET_ERR_MSG_WEAK_MOD(extack,
118 "error setting offload flag on port");
119 return err;
120 }
121
122 return 0;
123 }
124
br_switchdev_fdb_populate(struct net_bridge * br,struct switchdev_notifier_fdb_info * item,const struct net_bridge_fdb_entry * fdb,const void * ctx)125 static void br_switchdev_fdb_populate(struct net_bridge *br,
126 struct switchdev_notifier_fdb_info *item,
127 const struct net_bridge_fdb_entry *fdb,
128 const void *ctx)
129 {
130 const struct net_bridge_port *p = READ_ONCE(fdb->dst);
131
132 item->addr = fdb->key.addr.addr;
133 item->vid = fdb->key.vlan_id;
134 item->added_by_user = test_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
135 item->offloaded = test_bit(BR_FDB_OFFLOADED, &fdb->flags);
136 item->is_local = test_bit(BR_FDB_LOCAL, &fdb->flags);
137 item->locked = false;
138 item->info.dev = (!p || item->is_local) ? br->dev : p->dev;
139 item->info.ctx = ctx;
140 }
141
142 void
br_switchdev_fdb_notify(struct net_bridge * br,const struct net_bridge_fdb_entry * fdb,int type)143 br_switchdev_fdb_notify(struct net_bridge *br,
144 const struct net_bridge_fdb_entry *fdb, int type)
145 {
146 struct switchdev_notifier_fdb_info item;
147
148 if (test_bit(BR_FDB_LOCKED, &fdb->flags))
149 return;
150
151 br_switchdev_fdb_populate(br, &item, fdb, NULL);
152
153 switch (type) {
154 case RTM_DELNEIGH:
155 call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_DEVICE,
156 item.info.dev, &item.info, NULL);
157 break;
158 case RTM_NEWNEIGH:
159 call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_DEVICE,
160 item.info.dev, &item.info, NULL);
161 break;
162 }
163 }
164
br_switchdev_port_vlan_add(struct net_device * dev,u16 vid,u16 flags,bool changed,struct netlink_ext_ack * extack)165 int br_switchdev_port_vlan_add(struct net_device *dev, u16 vid, u16 flags,
166 bool changed, struct netlink_ext_ack *extack)
167 {
168 struct switchdev_obj_port_vlan v = {
169 .obj.orig_dev = dev,
170 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
171 .flags = flags,
172 .vid = vid,
173 .changed = changed,
174 };
175
176 return switchdev_port_obj_add(dev, &v.obj, extack);
177 }
178
br_switchdev_port_vlan_del(struct net_device * dev,u16 vid)179 int br_switchdev_port_vlan_del(struct net_device *dev, u16 vid)
180 {
181 struct switchdev_obj_port_vlan v = {
182 .obj.orig_dev = dev,
183 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
184 .vid = vid,
185 };
186
187 return switchdev_port_obj_del(dev, &v.obj);
188 }
189
nbp_switchdev_hwdom_set(struct net_bridge_port * joining)190 static int nbp_switchdev_hwdom_set(struct net_bridge_port *joining)
191 {
192 struct net_bridge *br = joining->br;
193 struct net_bridge_port *p;
194 int hwdom;
195
196 /* joining is yet to be added to the port list. */
197 list_for_each_entry(p, &br->port_list, list) {
198 if (netdev_phys_item_id_same(&joining->ppid, &p->ppid)) {
199 joining->hwdom = p->hwdom;
200 return 0;
201 }
202 }
203
204 hwdom = find_next_zero_bit(&br->busy_hwdoms, BR_HWDOM_MAX, 1);
205 if (hwdom >= BR_HWDOM_MAX)
206 return -EBUSY;
207
208 set_bit(hwdom, &br->busy_hwdoms);
209 joining->hwdom = hwdom;
210 return 0;
211 }
212
nbp_switchdev_hwdom_put(struct net_bridge_port * leaving)213 static void nbp_switchdev_hwdom_put(struct net_bridge_port *leaving)
214 {
215 struct net_bridge *br = leaving->br;
216 struct net_bridge_port *p;
217
218 /* leaving is no longer in the port list. */
219 list_for_each_entry(p, &br->port_list, list) {
220 if (p->hwdom == leaving->hwdom)
221 return;
222 }
223
224 clear_bit(leaving->hwdom, &br->busy_hwdoms);
225 }
226
nbp_switchdev_add(struct net_bridge_port * p,struct netdev_phys_item_id ppid,bool tx_fwd_offload,struct netlink_ext_ack * extack)227 static int nbp_switchdev_add(struct net_bridge_port *p,
228 struct netdev_phys_item_id ppid,
229 bool tx_fwd_offload,
230 struct netlink_ext_ack *extack)
231 {
232 int err;
233
234 if (p->offload_count) {
235 /* Prevent unsupported configurations such as a bridge port
236 * which is a bonding interface, and the member ports are from
237 * different hardware switches.
238 */
239 if (!netdev_phys_item_id_same(&p->ppid, &ppid)) {
240 NL_SET_ERR_MSG_MOD(extack,
241 "Same bridge port cannot be offloaded by two physical switches");
242 return -EBUSY;
243 }
244
245 /* Tolerate drivers that call switchdev_bridge_port_offload()
246 * more than once for the same bridge port, such as when the
247 * bridge port is an offloaded bonding/team interface.
248 */
249 p->offload_count++;
250
251 return 0;
252 }
253
254 p->ppid = ppid;
255 p->offload_count = 1;
256
257 err = nbp_switchdev_hwdom_set(p);
258 if (err)
259 return err;
260
261 if (tx_fwd_offload) {
262 p->flags |= BR_TX_FWD_OFFLOAD;
263 static_branch_inc(&br_switchdev_tx_fwd_offload);
264 }
265
266 return 0;
267 }
268
nbp_switchdev_del(struct net_bridge_port * p)269 static void nbp_switchdev_del(struct net_bridge_port *p)
270 {
271 if (WARN_ON(!p->offload_count))
272 return;
273
274 p->offload_count--;
275
276 if (p->offload_count)
277 return;
278
279 if (p->hwdom)
280 nbp_switchdev_hwdom_put(p);
281
282 if (p->flags & BR_TX_FWD_OFFLOAD) {
283 p->flags &= ~BR_TX_FWD_OFFLOAD;
284 static_branch_dec(&br_switchdev_tx_fwd_offload);
285 }
286 }
287
288 static int
br_switchdev_fdb_replay_one(struct net_bridge * br,struct notifier_block * nb,const struct net_bridge_fdb_entry * fdb,unsigned long action,const void * ctx)289 br_switchdev_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
290 const struct net_bridge_fdb_entry *fdb,
291 unsigned long action, const void *ctx)
292 {
293 struct switchdev_notifier_fdb_info item;
294 int err;
295
296 br_switchdev_fdb_populate(br, &item, fdb, ctx);
297
298 err = nb->notifier_call(nb, action, &item);
299 return notifier_to_errno(err);
300 }
301
302 static int
br_switchdev_fdb_replay(const struct net_device * br_dev,const void * ctx,bool adding,struct notifier_block * nb)303 br_switchdev_fdb_replay(const struct net_device *br_dev, const void *ctx,
304 bool adding, struct notifier_block *nb)
305 {
306 struct net_bridge_fdb_entry *fdb;
307 struct net_bridge *br;
308 unsigned long action;
309 int err = 0;
310
311 if (!nb)
312 return 0;
313
314 if (!netif_is_bridge_master(br_dev))
315 return -EINVAL;
316
317 br = netdev_priv(br_dev);
318
319 if (adding)
320 action = SWITCHDEV_FDB_ADD_TO_DEVICE;
321 else
322 action = SWITCHDEV_FDB_DEL_TO_DEVICE;
323
324 rcu_read_lock();
325
326 hlist_for_each_entry_rcu(fdb, &br->fdb_list, fdb_node) {
327 err = br_switchdev_fdb_replay_one(br, nb, fdb, action, ctx);
328 if (err)
329 break;
330 }
331
332 rcu_read_unlock();
333
334 return err;
335 }
336
br_switchdev_vlan_attr_replay(struct net_device * br_dev,const void * ctx,struct notifier_block * nb,struct netlink_ext_ack * extack)337 static int br_switchdev_vlan_attr_replay(struct net_device *br_dev,
338 const void *ctx,
339 struct notifier_block *nb,
340 struct netlink_ext_ack *extack)
341 {
342 struct switchdev_notifier_port_attr_info attr_info = {
343 .info = {
344 .dev = br_dev,
345 .extack = extack,
346 .ctx = ctx,
347 },
348 };
349 struct net_bridge *br = netdev_priv(br_dev);
350 struct net_bridge_vlan_group *vg;
351 struct switchdev_attr attr;
352 struct net_bridge_vlan *v;
353 int err;
354
355 attr_info.attr = &attr;
356 attr.orig_dev = br_dev;
357
358 vg = br_vlan_group(br);
359 if (!vg)
360 return 0;
361
362 list_for_each_entry(v, &vg->vlan_list, vlist) {
363 if (v->msti) {
364 attr.id = SWITCHDEV_ATTR_ID_VLAN_MSTI;
365 attr.u.vlan_msti.vid = v->vid;
366 attr.u.vlan_msti.msti = v->msti;
367
368 err = nb->notifier_call(nb, SWITCHDEV_PORT_ATTR_SET,
369 &attr_info);
370 err = notifier_to_errno(err);
371 if (err)
372 return err;
373 }
374 }
375
376 return 0;
377 }
378
379 static int
br_switchdev_vlan_replay_one(struct notifier_block * nb,struct net_device * dev,struct switchdev_obj_port_vlan * vlan,const void * ctx,unsigned long action,struct netlink_ext_ack * extack)380 br_switchdev_vlan_replay_one(struct notifier_block *nb,
381 struct net_device *dev,
382 struct switchdev_obj_port_vlan *vlan,
383 const void *ctx, unsigned long action,
384 struct netlink_ext_ack *extack)
385 {
386 struct switchdev_notifier_port_obj_info obj_info = {
387 .info = {
388 .dev = dev,
389 .extack = extack,
390 .ctx = ctx,
391 },
392 .obj = &vlan->obj,
393 };
394 int err;
395
396 err = nb->notifier_call(nb, action, &obj_info);
397 return notifier_to_errno(err);
398 }
399
br_switchdev_vlan_replay_group(struct notifier_block * nb,struct net_device * dev,struct net_bridge_vlan_group * vg,const void * ctx,unsigned long action,struct netlink_ext_ack * extack)400 static int br_switchdev_vlan_replay_group(struct notifier_block *nb,
401 struct net_device *dev,
402 struct net_bridge_vlan_group *vg,
403 const void *ctx, unsigned long action,
404 struct netlink_ext_ack *extack)
405 {
406 struct net_bridge_vlan *v;
407 int err = 0;
408 u16 pvid;
409
410 if (!vg)
411 return 0;
412
413 pvid = br_get_pvid(vg);
414
415 list_for_each_entry(v, &vg->vlan_list, vlist) {
416 struct switchdev_obj_port_vlan vlan = {
417 .obj.orig_dev = dev,
418 .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
419 .flags = br_vlan_flags(v, pvid),
420 .vid = v->vid,
421 };
422
423 if (!br_vlan_should_use(v))
424 continue;
425
426 err = br_switchdev_vlan_replay_one(nb, dev, &vlan, ctx,
427 action, extack);
428 if (err)
429 return err;
430 }
431
432 return 0;
433 }
434
br_switchdev_vlan_replay(struct net_device * br_dev,const void * ctx,bool adding,struct notifier_block * nb,struct netlink_ext_ack * extack)435 static int br_switchdev_vlan_replay(struct net_device *br_dev,
436 const void *ctx, bool adding,
437 struct notifier_block *nb,
438 struct netlink_ext_ack *extack)
439 {
440 struct net_bridge *br = netdev_priv(br_dev);
441 struct net_bridge_port *p;
442 unsigned long action;
443 int err;
444
445 ASSERT_RTNL();
446
447 if (!nb)
448 return 0;
449
450 if (!netif_is_bridge_master(br_dev))
451 return -EINVAL;
452
453 if (adding)
454 action = SWITCHDEV_PORT_OBJ_ADD;
455 else
456 action = SWITCHDEV_PORT_OBJ_DEL;
457
458 err = br_switchdev_vlan_replay_group(nb, br_dev, br_vlan_group(br),
459 ctx, action, extack);
460 if (err)
461 return err;
462
463 list_for_each_entry(p, &br->port_list, list) {
464 struct net_device *dev = p->dev;
465
466 err = br_switchdev_vlan_replay_group(nb, dev,
467 nbp_vlan_group(p),
468 ctx, action, extack);
469 if (err)
470 return err;
471 }
472
473 if (adding) {
474 err = br_switchdev_vlan_attr_replay(br_dev, ctx, nb, extack);
475 if (err)
476 return err;
477 }
478
479 return 0;
480 }
481
482 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
483 struct br_switchdev_mdb_complete_info {
484 struct net_bridge_port *port;
485 struct br_ip ip;
486 };
487
br_switchdev_mdb_complete(struct net_device * dev,int err,void * priv)488 static void br_switchdev_mdb_complete(struct net_device *dev, int err, void *priv)
489 {
490 struct br_switchdev_mdb_complete_info *data = priv;
491 struct net_bridge_port_group __rcu **pp;
492 struct net_bridge_port_group *p;
493 struct net_bridge_mdb_entry *mp;
494 struct net_bridge_port *port = data->port;
495 struct net_bridge *br = port->br;
496
497 if (err)
498 goto err;
499
500 spin_lock_bh(&br->multicast_lock);
501 mp = br_mdb_ip_get(br, &data->ip);
502 if (!mp)
503 goto out;
504 for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
505 pp = &p->next) {
506 if (p->key.port != port)
507 continue;
508 p->flags |= MDB_PG_FLAGS_OFFLOAD;
509 }
510 out:
511 spin_unlock_bh(&br->multicast_lock);
512 err:
513 kfree(priv);
514 }
515
br_switchdev_mdb_populate(struct switchdev_obj_port_mdb * mdb,const struct net_bridge_mdb_entry * mp)516 static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
517 const struct net_bridge_mdb_entry *mp)
518 {
519 if (mp->addr.proto == htons(ETH_P_IP))
520 ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
521 #if IS_ENABLED(CONFIG_IPV6)
522 else if (mp->addr.proto == htons(ETH_P_IPV6))
523 ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
524 #endif
525 else
526 ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
527
528 mdb->vid = mp->addr.vid;
529 }
530
br_switchdev_host_mdb_one(struct net_device * dev,struct net_device * lower_dev,struct net_bridge_mdb_entry * mp,int type)531 static void br_switchdev_host_mdb_one(struct net_device *dev,
532 struct net_device *lower_dev,
533 struct net_bridge_mdb_entry *mp,
534 int type)
535 {
536 struct switchdev_obj_port_mdb mdb = {
537 .obj = {
538 .id = SWITCHDEV_OBJ_ID_HOST_MDB,
539 .flags = SWITCHDEV_F_DEFER,
540 .orig_dev = dev,
541 },
542 };
543
544 br_switchdev_mdb_populate(&mdb, mp);
545
546 switch (type) {
547 case RTM_NEWMDB:
548 switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
549 break;
550 case RTM_DELMDB:
551 switchdev_port_obj_del(lower_dev, &mdb.obj);
552 break;
553 }
554 }
555
br_switchdev_host_mdb(struct net_device * dev,struct net_bridge_mdb_entry * mp,int type)556 static void br_switchdev_host_mdb(struct net_device *dev,
557 struct net_bridge_mdb_entry *mp, int type)
558 {
559 struct net_device *lower_dev;
560 struct list_head *iter;
561
562 netdev_for_each_lower_dev(dev, lower_dev, iter)
563 br_switchdev_host_mdb_one(dev, lower_dev, mp, type);
564 }
565
566 static int
br_switchdev_mdb_replay_one(struct notifier_block * nb,struct net_device * dev,const struct switchdev_obj_port_mdb * mdb,unsigned long action,const void * ctx,struct netlink_ext_ack * extack)567 br_switchdev_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
568 const struct switchdev_obj_port_mdb *mdb,
569 unsigned long action, const void *ctx,
570 struct netlink_ext_ack *extack)
571 {
572 struct switchdev_notifier_port_obj_info obj_info = {
573 .info = {
574 .dev = dev,
575 .extack = extack,
576 .ctx = ctx,
577 },
578 .obj = &mdb->obj,
579 };
580 int err;
581
582 err = nb->notifier_call(nb, action, &obj_info);
583 return notifier_to_errno(err);
584 }
585
br_switchdev_mdb_queue_one(struct list_head * mdb_list,enum switchdev_obj_id id,const struct net_bridge_mdb_entry * mp,struct net_device * orig_dev)586 static int br_switchdev_mdb_queue_one(struct list_head *mdb_list,
587 enum switchdev_obj_id id,
588 const struct net_bridge_mdb_entry *mp,
589 struct net_device *orig_dev)
590 {
591 struct switchdev_obj_port_mdb *mdb;
592
593 mdb = kzalloc(sizeof(*mdb), GFP_ATOMIC);
594 if (!mdb)
595 return -ENOMEM;
596
597 mdb->obj.id = id;
598 mdb->obj.orig_dev = orig_dev;
599 br_switchdev_mdb_populate(mdb, mp);
600 list_add_tail(&mdb->obj.list, mdb_list);
601
602 return 0;
603 }
604
br_switchdev_mdb_notify(struct net_device * dev,struct net_bridge_mdb_entry * mp,struct net_bridge_port_group * pg,int type)605 void br_switchdev_mdb_notify(struct net_device *dev,
606 struct net_bridge_mdb_entry *mp,
607 struct net_bridge_port_group *pg,
608 int type)
609 {
610 struct br_switchdev_mdb_complete_info *complete_info;
611 struct switchdev_obj_port_mdb mdb = {
612 .obj = {
613 .id = SWITCHDEV_OBJ_ID_PORT_MDB,
614 .flags = SWITCHDEV_F_DEFER,
615 },
616 };
617
618 if (!pg)
619 return br_switchdev_host_mdb(dev, mp, type);
620
621 br_switchdev_mdb_populate(&mdb, mp);
622
623 mdb.obj.orig_dev = pg->key.port->dev;
624 switch (type) {
625 case RTM_NEWMDB:
626 complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
627 if (!complete_info)
628 break;
629 complete_info->port = pg->key.port;
630 complete_info->ip = mp->addr;
631 mdb.obj.complete_priv = complete_info;
632 mdb.obj.complete = br_switchdev_mdb_complete;
633 if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
634 kfree(complete_info);
635 break;
636 case RTM_DELMDB:
637 switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
638 break;
639 }
640 }
641 #endif
642
643 static int
br_switchdev_mdb_replay(struct net_device * br_dev,struct net_device * dev,const void * ctx,bool adding,struct notifier_block * nb,struct netlink_ext_ack * extack)644 br_switchdev_mdb_replay(struct net_device *br_dev, struct net_device *dev,
645 const void *ctx, bool adding, struct notifier_block *nb,
646 struct netlink_ext_ack *extack)
647 {
648 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
649 const struct net_bridge_mdb_entry *mp;
650 struct switchdev_obj *obj, *tmp;
651 struct net_bridge *br;
652 unsigned long action;
653 LIST_HEAD(mdb_list);
654 int err = 0;
655
656 ASSERT_RTNL();
657
658 if (!nb)
659 return 0;
660
661 if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
662 return -EINVAL;
663
664 br = netdev_priv(br_dev);
665
666 if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
667 return 0;
668
669 /* We cannot walk over br->mdb_list protected just by the rtnl_mutex,
670 * because the write-side protection is br->multicast_lock. But we
671 * need to emulate the [ blocking ] calling context of a regular
672 * switchdev event, so since both br->multicast_lock and RCU read side
673 * critical sections are atomic, we have no choice but to pick the RCU
674 * read side lock, queue up all our events, leave the critical section
675 * and notify switchdev from blocking context.
676 */
677 rcu_read_lock();
678
679 hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
680 struct net_bridge_port_group __rcu * const *pp;
681 const struct net_bridge_port_group *p;
682
683 if (mp->host_joined) {
684 err = br_switchdev_mdb_queue_one(&mdb_list,
685 SWITCHDEV_OBJ_ID_HOST_MDB,
686 mp, br_dev);
687 if (err) {
688 rcu_read_unlock();
689 goto out_free_mdb;
690 }
691 }
692
693 for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
694 pp = &p->next) {
695 if (p->key.port->dev != dev)
696 continue;
697
698 err = br_switchdev_mdb_queue_one(&mdb_list,
699 SWITCHDEV_OBJ_ID_PORT_MDB,
700 mp, dev);
701 if (err) {
702 rcu_read_unlock();
703 goto out_free_mdb;
704 }
705 }
706 }
707
708 rcu_read_unlock();
709
710 if (adding)
711 action = SWITCHDEV_PORT_OBJ_ADD;
712 else
713 action = SWITCHDEV_PORT_OBJ_DEL;
714
715 list_for_each_entry(obj, &mdb_list, list) {
716 err = br_switchdev_mdb_replay_one(nb, dev,
717 SWITCHDEV_OBJ_PORT_MDB(obj),
718 action, ctx, extack);
719 if (err)
720 goto out_free_mdb;
721 }
722
723 out_free_mdb:
724 list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
725 list_del(&obj->list);
726 kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
727 }
728
729 if (err)
730 return err;
731 #endif
732
733 return 0;
734 }
735
nbp_switchdev_sync_objs(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb,struct netlink_ext_ack * extack)736 static int nbp_switchdev_sync_objs(struct net_bridge_port *p, const void *ctx,
737 struct notifier_block *atomic_nb,
738 struct notifier_block *blocking_nb,
739 struct netlink_ext_ack *extack)
740 {
741 struct net_device *br_dev = p->br->dev;
742 struct net_device *dev = p->dev;
743 int err;
744
745 err = br_switchdev_vlan_replay(br_dev, ctx, true, blocking_nb, extack);
746 if (err && err != -EOPNOTSUPP)
747 return err;
748
749 err = br_switchdev_mdb_replay(br_dev, dev, ctx, true, blocking_nb,
750 extack);
751 if (err && err != -EOPNOTSUPP)
752 return err;
753
754 err = br_switchdev_fdb_replay(br_dev, ctx, true, atomic_nb);
755 if (err && err != -EOPNOTSUPP)
756 return err;
757
758 return 0;
759 }
760
nbp_switchdev_unsync_objs(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb)761 static void nbp_switchdev_unsync_objs(struct net_bridge_port *p,
762 const void *ctx,
763 struct notifier_block *atomic_nb,
764 struct notifier_block *blocking_nb)
765 {
766 struct net_device *br_dev = p->br->dev;
767 struct net_device *dev = p->dev;
768
769 br_switchdev_fdb_replay(br_dev, ctx, false, atomic_nb);
770
771 br_switchdev_mdb_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
772
773 br_switchdev_vlan_replay(br_dev, ctx, false, blocking_nb, NULL);
774 }
775
776 /* Let the bridge know that this port is offloaded, so that it can assign a
777 * switchdev hardware domain to it.
778 */
br_switchdev_port_offload(struct net_bridge_port * p,struct net_device * dev,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb,bool tx_fwd_offload,struct netlink_ext_ack * extack)779 int br_switchdev_port_offload(struct net_bridge_port *p,
780 struct net_device *dev, const void *ctx,
781 struct notifier_block *atomic_nb,
782 struct notifier_block *blocking_nb,
783 bool tx_fwd_offload,
784 struct netlink_ext_ack *extack)
785 {
786 struct netdev_phys_item_id ppid;
787 int err;
788
789 err = dev_get_port_parent_id(dev, &ppid, false);
790 if (err)
791 return err;
792
793 err = nbp_switchdev_add(p, ppid, tx_fwd_offload, extack);
794 if (err)
795 return err;
796
797 err = nbp_switchdev_sync_objs(p, ctx, atomic_nb, blocking_nb, extack);
798 if (err)
799 goto out_switchdev_del;
800
801 return 0;
802
803 out_switchdev_del:
804 nbp_switchdev_del(p);
805
806 return err;
807 }
808
br_switchdev_port_unoffload(struct net_bridge_port * p,const void * ctx,struct notifier_block * atomic_nb,struct notifier_block * blocking_nb)809 void br_switchdev_port_unoffload(struct net_bridge_port *p, const void *ctx,
810 struct notifier_block *atomic_nb,
811 struct notifier_block *blocking_nb)
812 {
813 nbp_switchdev_unsync_objs(p, ctx, atomic_nb, blocking_nb);
814
815 nbp_switchdev_del(p);
816 }
817