Merge branch 'keys-asym-keyctl' into keys-next
[deliverable/linux.git] / net / bridge / br_netlink.c
1 /*
2 * Bridge netlink control interface
3 *
4 * Authors:
5 * Stephen Hemminger <shemminger@osdl.org>
6 *
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License
9 * as published by the Free Software Foundation; either version
10 * 2 of the License, or (at your option) any later version.
11 */
12
13 #include <linux/kernel.h>
14 #include <linux/slab.h>
15 #include <linux/etherdevice.h>
16 #include <net/rtnetlink.h>
17 #include <net/net_namespace.h>
18 #include <net/sock.h>
19 #include <uapi/linux/if_bridge.h>
20
21 #include "br_private.h"
22 #include "br_private_stp.h"
23
24 static int __get_num_vlan_infos(struct net_bridge_vlan_group *vg,
25 u32 filter_mask)
26 {
27 struct net_bridge_vlan *v;
28 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0;
29 u16 flags, pvid;
30 int num_vlans = 0;
31
32 if (!(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED))
33 return 0;
34
35 pvid = br_get_pvid(vg);
36 /* Count number of vlan infos */
37 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
38 flags = 0;
39 /* only a context, bridge vlan not activated */
40 if (!br_vlan_should_use(v))
41 continue;
42 if (v->vid == pvid)
43 flags |= BRIDGE_VLAN_INFO_PVID;
44
45 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
46 flags |= BRIDGE_VLAN_INFO_UNTAGGED;
47
48 if (vid_range_start == 0) {
49 goto initvars;
50 } else if ((v->vid - vid_range_end) == 1 &&
51 flags == vid_range_flags) {
52 vid_range_end = v->vid;
53 continue;
54 } else {
55 if ((vid_range_end - vid_range_start) > 0)
56 num_vlans += 2;
57 else
58 num_vlans += 1;
59 }
60 initvars:
61 vid_range_start = v->vid;
62 vid_range_end = v->vid;
63 vid_range_flags = flags;
64 }
65
66 if (vid_range_start != 0) {
67 if ((vid_range_end - vid_range_start) > 0)
68 num_vlans += 2;
69 else
70 num_vlans += 1;
71 }
72
73 return num_vlans;
74 }
75
76 static int br_get_num_vlan_infos(struct net_bridge_vlan_group *vg,
77 u32 filter_mask)
78 {
79 int num_vlans;
80
81 if (!vg)
82 return 0;
83
84 if (filter_mask & RTEXT_FILTER_BRVLAN)
85 return vg->num_vlans;
86
87 rcu_read_lock();
88 num_vlans = __get_num_vlan_infos(vg, filter_mask);
89 rcu_read_unlock();
90
91 return num_vlans;
92 }
93
94 static size_t br_get_link_af_size_filtered(const struct net_device *dev,
95 u32 filter_mask)
96 {
97 struct net_bridge_vlan_group *vg = NULL;
98 struct net_bridge_port *p;
99 struct net_bridge *br;
100 int num_vlan_infos;
101
102 rcu_read_lock();
103 if (br_port_exists(dev)) {
104 p = br_port_get_rcu(dev);
105 vg = nbp_vlan_group_rcu(p);
106 } else if (dev->priv_flags & IFF_EBRIDGE) {
107 br = netdev_priv(dev);
108 vg = br_vlan_group_rcu(br);
109 }
110 num_vlan_infos = br_get_num_vlan_infos(vg, filter_mask);
111 rcu_read_unlock();
112
113 /* Each VLAN is returned in bridge_vlan_info along with flags */
114 return num_vlan_infos * nla_total_size(sizeof(struct bridge_vlan_info));
115 }
116
117 static inline size_t br_port_info_size(void)
118 {
119 return nla_total_size(1) /* IFLA_BRPORT_STATE */
120 + nla_total_size(2) /* IFLA_BRPORT_PRIORITY */
121 + nla_total_size(4) /* IFLA_BRPORT_COST */
122 + nla_total_size(1) /* IFLA_BRPORT_MODE */
123 + nla_total_size(1) /* IFLA_BRPORT_GUARD */
124 + nla_total_size(1) /* IFLA_BRPORT_PROTECT */
125 + nla_total_size(1) /* IFLA_BRPORT_FAST_LEAVE */
126 + nla_total_size(1) /* IFLA_BRPORT_LEARNING */
127 + nla_total_size(1) /* IFLA_BRPORT_UNICAST_FLOOD */
128 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP */
129 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP_WIFI */
130 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_ROOT_ID */
131 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_BRIDGE_ID */
132 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_PORT */
133 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_COST */
134 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_ID */
135 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_NO */
136 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_TOPOLOGY_CHANGE_ACK */
137 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_CONFIG_PENDING */
138 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_MESSAGE_AGE_TIMER */
139 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_FORWARD_DELAY_TIMER */
140 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_HOLD_TIMER */
141 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
142 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_MULTICAST_ROUTER */
143 #endif
144 + 0;
145 }
146
147 static inline size_t br_nlmsg_size(struct net_device *dev, u32 filter_mask)
148 {
149 return NLMSG_ALIGN(sizeof(struct ifinfomsg))
150 + nla_total_size(IFNAMSIZ) /* IFLA_IFNAME */
151 + nla_total_size(MAX_ADDR_LEN) /* IFLA_ADDRESS */
152 + nla_total_size(4) /* IFLA_MASTER */
153 + nla_total_size(4) /* IFLA_MTU */
154 + nla_total_size(4) /* IFLA_LINK */
155 + nla_total_size(1) /* IFLA_OPERSTATE */
156 + nla_total_size(br_port_info_size()) /* IFLA_PROTINFO */
157 + nla_total_size(br_get_link_af_size_filtered(dev,
158 filter_mask)); /* IFLA_AF_SPEC */
159 }
160
161 static int br_port_fill_attrs(struct sk_buff *skb,
162 const struct net_bridge_port *p)
163 {
164 u8 mode = !!(p->flags & BR_HAIRPIN_MODE);
165 u64 timerval;
166
167 if (nla_put_u8(skb, IFLA_BRPORT_STATE, p->state) ||
168 nla_put_u16(skb, IFLA_BRPORT_PRIORITY, p->priority) ||
169 nla_put_u32(skb, IFLA_BRPORT_COST, p->path_cost) ||
170 nla_put_u8(skb, IFLA_BRPORT_MODE, mode) ||
171 nla_put_u8(skb, IFLA_BRPORT_GUARD, !!(p->flags & BR_BPDU_GUARD)) ||
172 nla_put_u8(skb, IFLA_BRPORT_PROTECT, !!(p->flags & BR_ROOT_BLOCK)) ||
173 nla_put_u8(skb, IFLA_BRPORT_FAST_LEAVE, !!(p->flags & BR_MULTICAST_FAST_LEAVE)) ||
174 nla_put_u8(skb, IFLA_BRPORT_LEARNING, !!(p->flags & BR_LEARNING)) ||
175 nla_put_u8(skb, IFLA_BRPORT_UNICAST_FLOOD, !!(p->flags & BR_FLOOD)) ||
176 nla_put_u8(skb, IFLA_BRPORT_PROXYARP, !!(p->flags & BR_PROXYARP)) ||
177 nla_put_u8(skb, IFLA_BRPORT_PROXYARP_WIFI,
178 !!(p->flags & BR_PROXYARP_WIFI)) ||
179 nla_put(skb, IFLA_BRPORT_ROOT_ID, sizeof(struct ifla_bridge_id),
180 &p->designated_root) ||
181 nla_put(skb, IFLA_BRPORT_BRIDGE_ID, sizeof(struct ifla_bridge_id),
182 &p->designated_bridge) ||
183 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_PORT, p->designated_port) ||
184 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_COST, p->designated_cost) ||
185 nla_put_u16(skb, IFLA_BRPORT_ID, p->port_id) ||
186 nla_put_u16(skb, IFLA_BRPORT_NO, p->port_no) ||
187 nla_put_u8(skb, IFLA_BRPORT_TOPOLOGY_CHANGE_ACK,
188 p->topology_change_ack) ||
189 nla_put_u8(skb, IFLA_BRPORT_CONFIG_PENDING, p->config_pending))
190 return -EMSGSIZE;
191
192 timerval = br_timer_value(&p->message_age_timer);
193 if (nla_put_u64_64bit(skb, IFLA_BRPORT_MESSAGE_AGE_TIMER, timerval,
194 IFLA_BRPORT_PAD))
195 return -EMSGSIZE;
196 timerval = br_timer_value(&p->forward_delay_timer);
197 if (nla_put_u64_64bit(skb, IFLA_BRPORT_FORWARD_DELAY_TIMER, timerval,
198 IFLA_BRPORT_PAD))
199 return -EMSGSIZE;
200 timerval = br_timer_value(&p->hold_timer);
201 if (nla_put_u64_64bit(skb, IFLA_BRPORT_HOLD_TIMER, timerval,
202 IFLA_BRPORT_PAD))
203 return -EMSGSIZE;
204
205 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
206 if (nla_put_u8(skb, IFLA_BRPORT_MULTICAST_ROUTER,
207 p->multicast_router))
208 return -EMSGSIZE;
209 #endif
210
211 return 0;
212 }
213
214 static int br_fill_ifvlaninfo_range(struct sk_buff *skb, u16 vid_start,
215 u16 vid_end, u16 flags)
216 {
217 struct bridge_vlan_info vinfo;
218
219 if ((vid_end - vid_start) > 0) {
220 /* add range to skb */
221 vinfo.vid = vid_start;
222 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_BEGIN;
223 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
224 sizeof(vinfo), &vinfo))
225 goto nla_put_failure;
226
227 vinfo.vid = vid_end;
228 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_END;
229 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
230 sizeof(vinfo), &vinfo))
231 goto nla_put_failure;
232 } else {
233 vinfo.vid = vid_start;
234 vinfo.flags = flags;
235 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
236 sizeof(vinfo), &vinfo))
237 goto nla_put_failure;
238 }
239
240 return 0;
241
242 nla_put_failure:
243 return -EMSGSIZE;
244 }
245
246 static int br_fill_ifvlaninfo_compressed(struct sk_buff *skb,
247 struct net_bridge_vlan_group *vg)
248 {
249 struct net_bridge_vlan *v;
250 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0;
251 u16 flags, pvid;
252 int err = 0;
253
254 /* Pack IFLA_BRIDGE_VLAN_INFO's for every vlan
255 * and mark vlan info with begin and end flags
256 * if vlaninfo represents a range
257 */
258 pvid = br_get_pvid(vg);
259 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
260 flags = 0;
261 if (!br_vlan_should_use(v))
262 continue;
263 if (v->vid == pvid)
264 flags |= BRIDGE_VLAN_INFO_PVID;
265
266 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
267 flags |= BRIDGE_VLAN_INFO_UNTAGGED;
268
269 if (vid_range_start == 0) {
270 goto initvars;
271 } else if ((v->vid - vid_range_end) == 1 &&
272 flags == vid_range_flags) {
273 vid_range_end = v->vid;
274 continue;
275 } else {
276 err = br_fill_ifvlaninfo_range(skb, vid_range_start,
277 vid_range_end,
278 vid_range_flags);
279 if (err)
280 return err;
281 }
282
283 initvars:
284 vid_range_start = v->vid;
285 vid_range_end = v->vid;
286 vid_range_flags = flags;
287 }
288
289 if (vid_range_start != 0) {
290 /* Call it once more to send any left over vlans */
291 err = br_fill_ifvlaninfo_range(skb, vid_range_start,
292 vid_range_end,
293 vid_range_flags);
294 if (err)
295 return err;
296 }
297
298 return 0;
299 }
300
301 static int br_fill_ifvlaninfo(struct sk_buff *skb,
302 struct net_bridge_vlan_group *vg)
303 {
304 struct bridge_vlan_info vinfo;
305 struct net_bridge_vlan *v;
306 u16 pvid;
307
308 pvid = br_get_pvid(vg);
309 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
310 if (!br_vlan_should_use(v))
311 continue;
312
313 vinfo.vid = v->vid;
314 vinfo.flags = 0;
315 if (v->vid == pvid)
316 vinfo.flags |= BRIDGE_VLAN_INFO_PVID;
317
318 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
319 vinfo.flags |= BRIDGE_VLAN_INFO_UNTAGGED;
320
321 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
322 sizeof(vinfo), &vinfo))
323 goto nla_put_failure;
324 }
325
326 return 0;
327
328 nla_put_failure:
329 return -EMSGSIZE;
330 }
331
332 /*
333 * Create one netlink message for one interface
334 * Contains port and master info as well as carrier and bridge state.
335 */
336 static int br_fill_ifinfo(struct sk_buff *skb,
337 struct net_bridge_port *port,
338 u32 pid, u32 seq, int event, unsigned int flags,
339 u32 filter_mask, const struct net_device *dev)
340 {
341 struct net_bridge *br;
342 struct ifinfomsg *hdr;
343 struct nlmsghdr *nlh;
344 u8 operstate = netif_running(dev) ? dev->operstate : IF_OPER_DOWN;
345
346 if (port)
347 br = port->br;
348 else
349 br = netdev_priv(dev);
350
351 br_debug(br, "br_fill_info event %d port %s master %s\n",
352 event, dev->name, br->dev->name);
353
354 nlh = nlmsg_put(skb, pid, seq, event, sizeof(*hdr), flags);
355 if (nlh == NULL)
356 return -EMSGSIZE;
357
358 hdr = nlmsg_data(nlh);
359 hdr->ifi_family = AF_BRIDGE;
360 hdr->__ifi_pad = 0;
361 hdr->ifi_type = dev->type;
362 hdr->ifi_index = dev->ifindex;
363 hdr->ifi_flags = dev_get_flags(dev);
364 hdr->ifi_change = 0;
365
366 if (nla_put_string(skb, IFLA_IFNAME, dev->name) ||
367 nla_put_u32(skb, IFLA_MASTER, br->dev->ifindex) ||
368 nla_put_u32(skb, IFLA_MTU, dev->mtu) ||
369 nla_put_u8(skb, IFLA_OPERSTATE, operstate) ||
370 (dev->addr_len &&
371 nla_put(skb, IFLA_ADDRESS, dev->addr_len, dev->dev_addr)) ||
372 (dev->ifindex != dev_get_iflink(dev) &&
373 nla_put_u32(skb, IFLA_LINK, dev_get_iflink(dev))))
374 goto nla_put_failure;
375
376 if (event == RTM_NEWLINK && port) {
377 struct nlattr *nest
378 = nla_nest_start(skb, IFLA_PROTINFO | NLA_F_NESTED);
379
380 if (nest == NULL || br_port_fill_attrs(skb, port) < 0)
381 goto nla_put_failure;
382 nla_nest_end(skb, nest);
383 }
384
385 /* Check if the VID information is requested */
386 if ((filter_mask & RTEXT_FILTER_BRVLAN) ||
387 (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)) {
388 struct net_bridge_vlan_group *vg;
389 struct nlattr *af;
390 int err;
391
392 /* RCU needed because of the VLAN locking rules (rcu || rtnl) */
393 rcu_read_lock();
394 if (port)
395 vg = nbp_vlan_group_rcu(port);
396 else
397 vg = br_vlan_group_rcu(br);
398
399 if (!vg || !vg->num_vlans) {
400 rcu_read_unlock();
401 goto done;
402 }
403 af = nla_nest_start(skb, IFLA_AF_SPEC);
404 if (!af) {
405 rcu_read_unlock();
406 goto nla_put_failure;
407 }
408 if (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)
409 err = br_fill_ifvlaninfo_compressed(skb, vg);
410 else
411 err = br_fill_ifvlaninfo(skb, vg);
412 rcu_read_unlock();
413 if (err)
414 goto nla_put_failure;
415 nla_nest_end(skb, af);
416 }
417
418 done:
419 nlmsg_end(skb, nlh);
420 return 0;
421
422 nla_put_failure:
423 nlmsg_cancel(skb, nlh);
424 return -EMSGSIZE;
425 }
426
427 /*
428 * Notify listeners of a change in port information
429 */
430 void br_ifinfo_notify(int event, struct net_bridge_port *port)
431 {
432 struct net *net;
433 struct sk_buff *skb;
434 int err = -ENOBUFS;
435 u32 filter = RTEXT_FILTER_BRVLAN_COMPRESSED;
436
437 if (!port)
438 return;
439
440 net = dev_net(port->dev);
441 br_debug(port->br, "port %u(%s) event %d\n",
442 (unsigned int)port->port_no, port->dev->name, event);
443
444 skb = nlmsg_new(br_nlmsg_size(port->dev, filter), GFP_ATOMIC);
445 if (skb == NULL)
446 goto errout;
447
448 err = br_fill_ifinfo(skb, port, 0, 0, event, 0, filter, port->dev);
449 if (err < 0) {
450 /* -EMSGSIZE implies BUG in br_nlmsg_size() */
451 WARN_ON(err == -EMSGSIZE);
452 kfree_skb(skb);
453 goto errout;
454 }
455 rtnl_notify(skb, net, 0, RTNLGRP_LINK, NULL, GFP_ATOMIC);
456 return;
457 errout:
458 rtnl_set_sk_err(net, RTNLGRP_LINK, err);
459 }
460
461
462 /*
463 * Dump information about all ports, in response to GETLINK
464 */
465 int br_getlink(struct sk_buff *skb, u32 pid, u32 seq,
466 struct net_device *dev, u32 filter_mask, int nlflags)
467 {
468 struct net_bridge_port *port = br_port_get_rtnl(dev);
469
470 if (!port && !(filter_mask & RTEXT_FILTER_BRVLAN) &&
471 !(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED))
472 return 0;
473
474 return br_fill_ifinfo(skb, port, pid, seq, RTM_NEWLINK, nlflags,
475 filter_mask, dev);
476 }
477
478 static int br_vlan_info(struct net_bridge *br, struct net_bridge_port *p,
479 int cmd, struct bridge_vlan_info *vinfo)
480 {
481 int err = 0;
482
483 switch (cmd) {
484 case RTM_SETLINK:
485 if (p) {
486 /* if the MASTER flag is set this will act on the global
487 * per-VLAN entry as well
488 */
489 err = nbp_vlan_add(p, vinfo->vid, vinfo->flags);
490 if (err)
491 break;
492 } else {
493 vinfo->flags |= BRIDGE_VLAN_INFO_BRENTRY;
494 err = br_vlan_add(br, vinfo->vid, vinfo->flags);
495 }
496 break;
497
498 case RTM_DELLINK:
499 if (p) {
500 nbp_vlan_delete(p, vinfo->vid);
501 if (vinfo->flags & BRIDGE_VLAN_INFO_MASTER)
502 br_vlan_delete(p->br, vinfo->vid);
503 } else {
504 br_vlan_delete(br, vinfo->vid);
505 }
506 break;
507 }
508
509 return err;
510 }
511
512 static int br_afspec(struct net_bridge *br,
513 struct net_bridge_port *p,
514 struct nlattr *af_spec,
515 int cmd)
516 {
517 struct bridge_vlan_info *vinfo_start = NULL;
518 struct bridge_vlan_info *vinfo = NULL;
519 struct nlattr *attr;
520 int err = 0;
521 int rem;
522
523 nla_for_each_nested(attr, af_spec, rem) {
524 if (nla_type(attr) != IFLA_BRIDGE_VLAN_INFO)
525 continue;
526 if (nla_len(attr) != sizeof(struct bridge_vlan_info))
527 return -EINVAL;
528 vinfo = nla_data(attr);
529 if (!vinfo->vid || vinfo->vid >= VLAN_VID_MASK)
530 return -EINVAL;
531 if (vinfo->flags & BRIDGE_VLAN_INFO_RANGE_BEGIN) {
532 if (vinfo_start)
533 return -EINVAL;
534 vinfo_start = vinfo;
535 /* don't allow range of pvids */
536 if (vinfo_start->flags & BRIDGE_VLAN_INFO_PVID)
537 return -EINVAL;
538 continue;
539 }
540
541 if (vinfo_start) {
542 struct bridge_vlan_info tmp_vinfo;
543 int v;
544
545 if (!(vinfo->flags & BRIDGE_VLAN_INFO_RANGE_END))
546 return -EINVAL;
547
548 if (vinfo->vid <= vinfo_start->vid)
549 return -EINVAL;
550
551 memcpy(&tmp_vinfo, vinfo_start,
552 sizeof(struct bridge_vlan_info));
553
554 for (v = vinfo_start->vid; v <= vinfo->vid; v++) {
555 tmp_vinfo.vid = v;
556 err = br_vlan_info(br, p, cmd, &tmp_vinfo);
557 if (err)
558 break;
559 }
560 vinfo_start = NULL;
561 } else {
562 err = br_vlan_info(br, p, cmd, vinfo);
563 }
564 if (err)
565 break;
566 }
567
568 return err;
569 }
570
571 static const struct nla_policy br_port_policy[IFLA_BRPORT_MAX + 1] = {
572 [IFLA_BRPORT_STATE] = { .type = NLA_U8 },
573 [IFLA_BRPORT_COST] = { .type = NLA_U32 },
574 [IFLA_BRPORT_PRIORITY] = { .type = NLA_U16 },
575 [IFLA_BRPORT_MODE] = { .type = NLA_U8 },
576 [IFLA_BRPORT_GUARD] = { .type = NLA_U8 },
577 [IFLA_BRPORT_PROTECT] = { .type = NLA_U8 },
578 [IFLA_BRPORT_FAST_LEAVE]= { .type = NLA_U8 },
579 [IFLA_BRPORT_LEARNING] = { .type = NLA_U8 },
580 [IFLA_BRPORT_UNICAST_FLOOD] = { .type = NLA_U8 },
581 [IFLA_BRPORT_PROXYARP] = { .type = NLA_U8 },
582 [IFLA_BRPORT_PROXYARP_WIFI] = { .type = NLA_U8 },
583 [IFLA_BRPORT_MULTICAST_ROUTER] = { .type = NLA_U8 },
584 };
585
586 /* Change the state of the port and notify spanning tree */
587 static int br_set_port_state(struct net_bridge_port *p, u8 state)
588 {
589 if (state > BR_STATE_BLOCKING)
590 return -EINVAL;
591
592 /* if kernel STP is running, don't allow changes */
593 if (p->br->stp_enabled == BR_KERNEL_STP)
594 return -EBUSY;
595
596 /* if device is not up, change is not allowed
597 * if link is not present, only allowable state is disabled
598 */
599 if (!netif_running(p->dev) ||
600 (!netif_oper_up(p->dev) && state != BR_STATE_DISABLED))
601 return -ENETDOWN;
602
603 br_set_state(p, state);
604 br_port_state_selection(p->br);
605 return 0;
606 }
607
608 /* Set/clear or port flags based on attribute */
609 static void br_set_port_flag(struct net_bridge_port *p, struct nlattr *tb[],
610 int attrtype, unsigned long mask)
611 {
612 if (tb[attrtype]) {
613 u8 flag = nla_get_u8(tb[attrtype]);
614 if (flag)
615 p->flags |= mask;
616 else
617 p->flags &= ~mask;
618 }
619 }
620
621 /* Process bridge protocol info on port */
622 static int br_setport(struct net_bridge_port *p, struct nlattr *tb[])
623 {
624 int err;
625 unsigned long old_flags = p->flags;
626
627 br_set_port_flag(p, tb, IFLA_BRPORT_MODE, BR_HAIRPIN_MODE);
628 br_set_port_flag(p, tb, IFLA_BRPORT_GUARD, BR_BPDU_GUARD);
629 br_set_port_flag(p, tb, IFLA_BRPORT_FAST_LEAVE, BR_MULTICAST_FAST_LEAVE);
630 br_set_port_flag(p, tb, IFLA_BRPORT_PROTECT, BR_ROOT_BLOCK);
631 br_set_port_flag(p, tb, IFLA_BRPORT_LEARNING, BR_LEARNING);
632 br_set_port_flag(p, tb, IFLA_BRPORT_UNICAST_FLOOD, BR_FLOOD);
633 br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP, BR_PROXYARP);
634 br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP_WIFI, BR_PROXYARP_WIFI);
635
636 if (tb[IFLA_BRPORT_COST]) {
637 err = br_stp_set_path_cost(p, nla_get_u32(tb[IFLA_BRPORT_COST]));
638 if (err)
639 return err;
640 }
641
642 if (tb[IFLA_BRPORT_PRIORITY]) {
643 err = br_stp_set_port_priority(p, nla_get_u16(tb[IFLA_BRPORT_PRIORITY]));
644 if (err)
645 return err;
646 }
647
648 if (tb[IFLA_BRPORT_STATE]) {
649 err = br_set_port_state(p, nla_get_u8(tb[IFLA_BRPORT_STATE]));
650 if (err)
651 return err;
652 }
653
654 if (tb[IFLA_BRPORT_FLUSH])
655 br_fdb_delete_by_port(p->br, p, 0, 0);
656
657 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
658 if (tb[IFLA_BRPORT_MULTICAST_ROUTER]) {
659 u8 mcast_router = nla_get_u8(tb[IFLA_BRPORT_MULTICAST_ROUTER]);
660
661 err = br_multicast_set_port_router(p, mcast_router);
662 if (err)
663 return err;
664 }
665 #endif
666 br_port_flags_change(p, old_flags ^ p->flags);
667 return 0;
668 }
669
670 /* Change state and parameters on port. */
671 int br_setlink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags)
672 {
673 struct nlattr *protinfo;
674 struct nlattr *afspec;
675 struct net_bridge_port *p;
676 struct nlattr *tb[IFLA_BRPORT_MAX + 1];
677 int err = 0;
678
679 protinfo = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_PROTINFO);
680 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
681 if (!protinfo && !afspec)
682 return 0;
683
684 p = br_port_get_rtnl(dev);
685 /* We want to accept dev as bridge itself if the AF_SPEC
686 * is set to see if someone is setting vlan info on the bridge
687 */
688 if (!p && !afspec)
689 return -EINVAL;
690
691 if (p && protinfo) {
692 if (protinfo->nla_type & NLA_F_NESTED) {
693 err = nla_parse_nested(tb, IFLA_BRPORT_MAX,
694 protinfo, br_port_policy);
695 if (err)
696 return err;
697
698 spin_lock_bh(&p->br->lock);
699 err = br_setport(p, tb);
700 spin_unlock_bh(&p->br->lock);
701 } else {
702 /* Binary compatibility with old RSTP */
703 if (nla_len(protinfo) < sizeof(u8))
704 return -EINVAL;
705
706 spin_lock_bh(&p->br->lock);
707 err = br_set_port_state(p, nla_get_u8(protinfo));
708 spin_unlock_bh(&p->br->lock);
709 }
710 if (err)
711 goto out;
712 }
713
714 if (afspec) {
715 err = br_afspec((struct net_bridge *)netdev_priv(dev), p,
716 afspec, RTM_SETLINK);
717 }
718
719 if (err == 0)
720 br_ifinfo_notify(RTM_NEWLINK, p);
721 out:
722 return err;
723 }
724
725 /* Delete port information */
726 int br_dellink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags)
727 {
728 struct nlattr *afspec;
729 struct net_bridge_port *p;
730 int err = 0;
731
732 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
733 if (!afspec)
734 return 0;
735
736 p = br_port_get_rtnl(dev);
737 /* We want to accept dev as bridge itself as well */
738 if (!p && !(dev->priv_flags & IFF_EBRIDGE))
739 return -EINVAL;
740
741 err = br_afspec((struct net_bridge *)netdev_priv(dev), p,
742 afspec, RTM_DELLINK);
743 if (err == 0)
744 /* Send RTM_NEWLINK because userspace
745 * expects RTM_NEWLINK for vlan dels
746 */
747 br_ifinfo_notify(RTM_NEWLINK, p);
748
749 return err;
750 }
751 static int br_validate(struct nlattr *tb[], struct nlattr *data[])
752 {
753 if (tb[IFLA_ADDRESS]) {
754 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
755 return -EINVAL;
756 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
757 return -EADDRNOTAVAIL;
758 }
759
760 if (!data)
761 return 0;
762
763 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
764 if (data[IFLA_BR_VLAN_PROTOCOL]) {
765 switch (nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL])) {
766 case htons(ETH_P_8021Q):
767 case htons(ETH_P_8021AD):
768 break;
769 default:
770 return -EPROTONOSUPPORT;
771 }
772 }
773 #endif
774
775 return 0;
776 }
777
778 static int br_dev_newlink(struct net *src_net, struct net_device *dev,
779 struct nlattr *tb[], struct nlattr *data[])
780 {
781 struct net_bridge *br = netdev_priv(dev);
782
783 if (tb[IFLA_ADDRESS]) {
784 spin_lock_bh(&br->lock);
785 br_stp_change_bridge_id(br, nla_data(tb[IFLA_ADDRESS]));
786 spin_unlock_bh(&br->lock);
787 }
788
789 return register_netdevice(dev);
790 }
791
792 static int br_port_slave_changelink(struct net_device *brdev,
793 struct net_device *dev,
794 struct nlattr *tb[],
795 struct nlattr *data[])
796 {
797 struct net_bridge *br = netdev_priv(brdev);
798 int ret;
799
800 if (!data)
801 return 0;
802
803 spin_lock_bh(&br->lock);
804 ret = br_setport(br_port_get_rtnl(dev), data);
805 spin_unlock_bh(&br->lock);
806
807 return ret;
808 }
809
810 static int br_port_fill_slave_info(struct sk_buff *skb,
811 const struct net_device *brdev,
812 const struct net_device *dev)
813 {
814 return br_port_fill_attrs(skb, br_port_get_rtnl(dev));
815 }
816
817 static size_t br_port_get_slave_size(const struct net_device *brdev,
818 const struct net_device *dev)
819 {
820 return br_port_info_size();
821 }
822
823 static const struct nla_policy br_policy[IFLA_BR_MAX + 1] = {
824 [IFLA_BR_FORWARD_DELAY] = { .type = NLA_U32 },
825 [IFLA_BR_HELLO_TIME] = { .type = NLA_U32 },
826 [IFLA_BR_MAX_AGE] = { .type = NLA_U32 },
827 [IFLA_BR_AGEING_TIME] = { .type = NLA_U32 },
828 [IFLA_BR_STP_STATE] = { .type = NLA_U32 },
829 [IFLA_BR_PRIORITY] = { .type = NLA_U16 },
830 [IFLA_BR_VLAN_FILTERING] = { .type = NLA_U8 },
831 [IFLA_BR_VLAN_PROTOCOL] = { .type = NLA_U16 },
832 [IFLA_BR_GROUP_FWD_MASK] = { .type = NLA_U16 },
833 [IFLA_BR_GROUP_ADDR] = { .type = NLA_BINARY,
834 .len = ETH_ALEN },
835 [IFLA_BR_MCAST_ROUTER] = { .type = NLA_U8 },
836 [IFLA_BR_MCAST_SNOOPING] = { .type = NLA_U8 },
837 [IFLA_BR_MCAST_QUERY_USE_IFADDR] = { .type = NLA_U8 },
838 [IFLA_BR_MCAST_QUERIER] = { .type = NLA_U8 },
839 [IFLA_BR_MCAST_HASH_ELASTICITY] = { .type = NLA_U32 },
840 [IFLA_BR_MCAST_HASH_MAX] = { .type = NLA_U32 },
841 [IFLA_BR_MCAST_LAST_MEMBER_CNT] = { .type = NLA_U32 },
842 [IFLA_BR_MCAST_STARTUP_QUERY_CNT] = { .type = NLA_U32 },
843 [IFLA_BR_MCAST_LAST_MEMBER_INTVL] = { .type = NLA_U64 },
844 [IFLA_BR_MCAST_MEMBERSHIP_INTVL] = { .type = NLA_U64 },
845 [IFLA_BR_MCAST_QUERIER_INTVL] = { .type = NLA_U64 },
846 [IFLA_BR_MCAST_QUERY_INTVL] = { .type = NLA_U64 },
847 [IFLA_BR_MCAST_QUERY_RESPONSE_INTVL] = { .type = NLA_U64 },
848 [IFLA_BR_MCAST_STARTUP_QUERY_INTVL] = { .type = NLA_U64 },
849 [IFLA_BR_NF_CALL_IPTABLES] = { .type = NLA_U8 },
850 [IFLA_BR_NF_CALL_IP6TABLES] = { .type = NLA_U8 },
851 [IFLA_BR_NF_CALL_ARPTABLES] = { .type = NLA_U8 },
852 [IFLA_BR_VLAN_DEFAULT_PVID] = { .type = NLA_U16 },
853 [IFLA_BR_VLAN_STATS_ENABLED] = { .type = NLA_U8 },
854 [IFLA_BR_MCAST_STATS_ENABLED] = { .type = NLA_U8 },
855 };
856
857 static int br_changelink(struct net_device *brdev, struct nlattr *tb[],
858 struct nlattr *data[])
859 {
860 struct net_bridge *br = netdev_priv(brdev);
861 int err;
862
863 if (!data)
864 return 0;
865
866 if (data[IFLA_BR_FORWARD_DELAY]) {
867 err = br_set_forward_delay(br, nla_get_u32(data[IFLA_BR_FORWARD_DELAY]));
868 if (err)
869 return err;
870 }
871
872 if (data[IFLA_BR_HELLO_TIME]) {
873 err = br_set_hello_time(br, nla_get_u32(data[IFLA_BR_HELLO_TIME]));
874 if (err)
875 return err;
876 }
877
878 if (data[IFLA_BR_MAX_AGE]) {
879 err = br_set_max_age(br, nla_get_u32(data[IFLA_BR_MAX_AGE]));
880 if (err)
881 return err;
882 }
883
884 if (data[IFLA_BR_AGEING_TIME]) {
885 err = br_set_ageing_time(br, nla_get_u32(data[IFLA_BR_AGEING_TIME]));
886 if (err)
887 return err;
888 }
889
890 if (data[IFLA_BR_STP_STATE]) {
891 u32 stp_enabled = nla_get_u32(data[IFLA_BR_STP_STATE]);
892
893 br_stp_set_enabled(br, stp_enabled);
894 }
895
896 if (data[IFLA_BR_PRIORITY]) {
897 u32 priority = nla_get_u16(data[IFLA_BR_PRIORITY]);
898
899 br_stp_set_bridge_priority(br, priority);
900 }
901
902 if (data[IFLA_BR_VLAN_FILTERING]) {
903 u8 vlan_filter = nla_get_u8(data[IFLA_BR_VLAN_FILTERING]);
904
905 err = __br_vlan_filter_toggle(br, vlan_filter);
906 if (err)
907 return err;
908 }
909
910 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
911 if (data[IFLA_BR_VLAN_PROTOCOL]) {
912 __be16 vlan_proto = nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL]);
913
914 err = __br_vlan_set_proto(br, vlan_proto);
915 if (err)
916 return err;
917 }
918
919 if (data[IFLA_BR_VLAN_DEFAULT_PVID]) {
920 __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]);
921
922 err = __br_vlan_set_default_pvid(br, defpvid);
923 if (err)
924 return err;
925 }
926
927 if (data[IFLA_BR_VLAN_STATS_ENABLED]) {
928 __u8 vlan_stats = nla_get_u8(data[IFLA_BR_VLAN_STATS_ENABLED]);
929
930 err = br_vlan_set_stats(br, vlan_stats);
931 if (err)
932 return err;
933 }
934 #endif
935
936 if (data[IFLA_BR_GROUP_FWD_MASK]) {
937 u16 fwd_mask = nla_get_u16(data[IFLA_BR_GROUP_FWD_MASK]);
938
939 if (fwd_mask & BR_GROUPFWD_RESTRICTED)
940 return -EINVAL;
941 br->group_fwd_mask = fwd_mask;
942 }
943
944 if (data[IFLA_BR_GROUP_ADDR]) {
945 u8 new_addr[ETH_ALEN];
946
947 if (nla_len(data[IFLA_BR_GROUP_ADDR]) != ETH_ALEN)
948 return -EINVAL;
949 memcpy(new_addr, nla_data(data[IFLA_BR_GROUP_ADDR]), ETH_ALEN);
950 if (!is_link_local_ether_addr(new_addr))
951 return -EINVAL;
952 if (new_addr[5] == 1 || /* 802.3x Pause address */
953 new_addr[5] == 2 || /* 802.3ad Slow protocols */
954 new_addr[5] == 3) /* 802.1X PAE address */
955 return -EINVAL;
956 spin_lock_bh(&br->lock);
957 memcpy(br->group_addr, new_addr, sizeof(br->group_addr));
958 spin_unlock_bh(&br->lock);
959 br->group_addr_set = true;
960 br_recalculate_fwd_mask(br);
961 }
962
963 if (data[IFLA_BR_FDB_FLUSH])
964 br_fdb_flush(br);
965
966 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
967 if (data[IFLA_BR_MCAST_ROUTER]) {
968 u8 multicast_router = nla_get_u8(data[IFLA_BR_MCAST_ROUTER]);
969
970 err = br_multicast_set_router(br, multicast_router);
971 if (err)
972 return err;
973 }
974
975 if (data[IFLA_BR_MCAST_SNOOPING]) {
976 u8 mcast_snooping = nla_get_u8(data[IFLA_BR_MCAST_SNOOPING]);
977
978 err = br_multicast_toggle(br, mcast_snooping);
979 if (err)
980 return err;
981 }
982
983 if (data[IFLA_BR_MCAST_QUERY_USE_IFADDR]) {
984 u8 val;
985
986 val = nla_get_u8(data[IFLA_BR_MCAST_QUERY_USE_IFADDR]);
987 br->multicast_query_use_ifaddr = !!val;
988 }
989
990 if (data[IFLA_BR_MCAST_QUERIER]) {
991 u8 mcast_querier = nla_get_u8(data[IFLA_BR_MCAST_QUERIER]);
992
993 err = br_multicast_set_querier(br, mcast_querier);
994 if (err)
995 return err;
996 }
997
998 if (data[IFLA_BR_MCAST_HASH_ELASTICITY]) {
999 u32 val = nla_get_u32(data[IFLA_BR_MCAST_HASH_ELASTICITY]);
1000
1001 br->hash_elasticity = val;
1002 }
1003
1004 if (data[IFLA_BR_MCAST_HASH_MAX]) {
1005 u32 hash_max = nla_get_u32(data[IFLA_BR_MCAST_HASH_MAX]);
1006
1007 err = br_multicast_set_hash_max(br, hash_max);
1008 if (err)
1009 return err;
1010 }
1011
1012 if (data[IFLA_BR_MCAST_LAST_MEMBER_CNT]) {
1013 u32 val = nla_get_u32(data[IFLA_BR_MCAST_LAST_MEMBER_CNT]);
1014
1015 br->multicast_last_member_count = val;
1016 }
1017
1018 if (data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]) {
1019 u32 val = nla_get_u32(data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]);
1020
1021 br->multicast_startup_query_count = val;
1022 }
1023
1024 if (data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]) {
1025 u64 val = nla_get_u64(data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]);
1026
1027 br->multicast_last_member_interval = clock_t_to_jiffies(val);
1028 }
1029
1030 if (data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]) {
1031 u64 val = nla_get_u64(data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]);
1032
1033 br->multicast_membership_interval = clock_t_to_jiffies(val);
1034 }
1035
1036 if (data[IFLA_BR_MCAST_QUERIER_INTVL]) {
1037 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERIER_INTVL]);
1038
1039 br->multicast_querier_interval = clock_t_to_jiffies(val);
1040 }
1041
1042 if (data[IFLA_BR_MCAST_QUERY_INTVL]) {
1043 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_INTVL]);
1044
1045 br->multicast_query_interval = clock_t_to_jiffies(val);
1046 }
1047
1048 if (data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]) {
1049 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]);
1050
1051 br->multicast_query_response_interval = clock_t_to_jiffies(val);
1052 }
1053
1054 if (data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]) {
1055 u64 val = nla_get_u64(data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]);
1056
1057 br->multicast_startup_query_interval = clock_t_to_jiffies(val);
1058 }
1059
1060 if (data[IFLA_BR_MCAST_STATS_ENABLED]) {
1061 __u8 mcast_stats;
1062
1063 mcast_stats = nla_get_u8(data[IFLA_BR_MCAST_STATS_ENABLED]);
1064 br->multicast_stats_enabled = !!mcast_stats;
1065 }
1066 #endif
1067 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1068 if (data[IFLA_BR_NF_CALL_IPTABLES]) {
1069 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IPTABLES]);
1070
1071 br->nf_call_iptables = val ? true : false;
1072 }
1073
1074 if (data[IFLA_BR_NF_CALL_IP6TABLES]) {
1075 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IP6TABLES]);
1076
1077 br->nf_call_ip6tables = val ? true : false;
1078 }
1079
1080 if (data[IFLA_BR_NF_CALL_ARPTABLES]) {
1081 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_ARPTABLES]);
1082
1083 br->nf_call_arptables = val ? true : false;
1084 }
1085 #endif
1086
1087 return 0;
1088 }
1089
1090 static size_t br_get_size(const struct net_device *brdev)
1091 {
1092 return nla_total_size(sizeof(u32)) + /* IFLA_BR_FORWARD_DELAY */
1093 nla_total_size(sizeof(u32)) + /* IFLA_BR_HELLO_TIME */
1094 nla_total_size(sizeof(u32)) + /* IFLA_BR_MAX_AGE */
1095 nla_total_size(sizeof(u32)) + /* IFLA_BR_AGEING_TIME */
1096 nla_total_size(sizeof(u32)) + /* IFLA_BR_STP_STATE */
1097 nla_total_size(sizeof(u16)) + /* IFLA_BR_PRIORITY */
1098 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_FILTERING */
1099 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
1100 nla_total_size(sizeof(__be16)) + /* IFLA_BR_VLAN_PROTOCOL */
1101 nla_total_size(sizeof(u16)) + /* IFLA_BR_VLAN_DEFAULT_PVID */
1102 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_STATS_ENABLED */
1103 #endif
1104 nla_total_size(sizeof(u16)) + /* IFLA_BR_GROUP_FWD_MASK */
1105 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_ROOT_ID */
1106 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_BRIDGE_ID */
1107 nla_total_size(sizeof(u16)) + /* IFLA_BR_ROOT_PORT */
1108 nla_total_size(sizeof(u32)) + /* IFLA_BR_ROOT_PATH_COST */
1109 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE */
1110 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE_DETECTED */
1111 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_HELLO_TIMER */
1112 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TCN_TIMER */
1113 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TOPOLOGY_CHANGE_TIMER */
1114 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_GC_TIMER */
1115 nla_total_size(ETH_ALEN) + /* IFLA_BR_GROUP_ADDR */
1116 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1117 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_ROUTER */
1118 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_SNOOPING */
1119 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERY_USE_IFADDR */
1120 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERIER */
1121 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_STATS_ENABLED */
1122 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_ELASTICITY */
1123 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_MAX */
1124 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_LAST_MEMBER_CNT */
1125 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_STARTUP_QUERY_CNT */
1126 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_LAST_MEMBER_INTVL */
1127 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_MEMBERSHIP_INTVL */
1128 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERIER_INTVL */
1129 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_INTVL */
1130 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_RESPONSE_INTVL */
1131 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_STARTUP_QUERY_INTVL */
1132 #endif
1133 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1134 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IPTABLES */
1135 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IP6TABLES */
1136 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_ARPTABLES */
1137 #endif
1138 0;
1139 }
1140
1141 static int br_fill_info(struct sk_buff *skb, const struct net_device *brdev)
1142 {
1143 struct net_bridge *br = netdev_priv(brdev);
1144 u32 forward_delay = jiffies_to_clock_t(br->forward_delay);
1145 u32 hello_time = jiffies_to_clock_t(br->hello_time);
1146 u32 age_time = jiffies_to_clock_t(br->max_age);
1147 u32 ageing_time = jiffies_to_clock_t(br->ageing_time);
1148 u32 stp_enabled = br->stp_enabled;
1149 u16 priority = (br->bridge_id.prio[0] << 8) | br->bridge_id.prio[1];
1150 u8 vlan_enabled = br_vlan_enabled(br);
1151 u64 clockval;
1152
1153 clockval = br_timer_value(&br->hello_timer);
1154 if (nla_put_u64_64bit(skb, IFLA_BR_HELLO_TIMER, clockval, IFLA_BR_PAD))
1155 return -EMSGSIZE;
1156 clockval = br_timer_value(&br->tcn_timer);
1157 if (nla_put_u64_64bit(skb, IFLA_BR_TCN_TIMER, clockval, IFLA_BR_PAD))
1158 return -EMSGSIZE;
1159 clockval = br_timer_value(&br->topology_change_timer);
1160 if (nla_put_u64_64bit(skb, IFLA_BR_TOPOLOGY_CHANGE_TIMER, clockval,
1161 IFLA_BR_PAD))
1162 return -EMSGSIZE;
1163 clockval = br_timer_value(&br->gc_timer);
1164 if (nla_put_u64_64bit(skb, IFLA_BR_GC_TIMER, clockval, IFLA_BR_PAD))
1165 return -EMSGSIZE;
1166
1167 if (nla_put_u32(skb, IFLA_BR_FORWARD_DELAY, forward_delay) ||
1168 nla_put_u32(skb, IFLA_BR_HELLO_TIME, hello_time) ||
1169 nla_put_u32(skb, IFLA_BR_MAX_AGE, age_time) ||
1170 nla_put_u32(skb, IFLA_BR_AGEING_TIME, ageing_time) ||
1171 nla_put_u32(skb, IFLA_BR_STP_STATE, stp_enabled) ||
1172 nla_put_u16(skb, IFLA_BR_PRIORITY, priority) ||
1173 nla_put_u8(skb, IFLA_BR_VLAN_FILTERING, vlan_enabled) ||
1174 nla_put_u16(skb, IFLA_BR_GROUP_FWD_MASK, br->group_fwd_mask) ||
1175 nla_put(skb, IFLA_BR_BRIDGE_ID, sizeof(struct ifla_bridge_id),
1176 &br->bridge_id) ||
1177 nla_put(skb, IFLA_BR_ROOT_ID, sizeof(struct ifla_bridge_id),
1178 &br->designated_root) ||
1179 nla_put_u16(skb, IFLA_BR_ROOT_PORT, br->root_port) ||
1180 nla_put_u32(skb, IFLA_BR_ROOT_PATH_COST, br->root_path_cost) ||
1181 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE, br->topology_change) ||
1182 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE_DETECTED,
1183 br->topology_change_detected) ||
1184 nla_put(skb, IFLA_BR_GROUP_ADDR, ETH_ALEN, br->group_addr))
1185 return -EMSGSIZE;
1186
1187 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
1188 if (nla_put_be16(skb, IFLA_BR_VLAN_PROTOCOL, br->vlan_proto) ||
1189 nla_put_u16(skb, IFLA_BR_VLAN_DEFAULT_PVID, br->default_pvid) ||
1190 nla_put_u8(skb, IFLA_BR_VLAN_STATS_ENABLED, br->vlan_stats_enabled))
1191 return -EMSGSIZE;
1192 #endif
1193 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1194 if (nla_put_u8(skb, IFLA_BR_MCAST_ROUTER, br->multicast_router) ||
1195 nla_put_u8(skb, IFLA_BR_MCAST_SNOOPING, !br->multicast_disabled) ||
1196 nla_put_u8(skb, IFLA_BR_MCAST_QUERY_USE_IFADDR,
1197 br->multicast_query_use_ifaddr) ||
1198 nla_put_u8(skb, IFLA_BR_MCAST_QUERIER, br->multicast_querier) ||
1199 nla_put_u8(skb, IFLA_BR_MCAST_STATS_ENABLED,
1200 br->multicast_stats_enabled) ||
1201 nla_put_u32(skb, IFLA_BR_MCAST_HASH_ELASTICITY,
1202 br->hash_elasticity) ||
1203 nla_put_u32(skb, IFLA_BR_MCAST_HASH_MAX, br->hash_max) ||
1204 nla_put_u32(skb, IFLA_BR_MCAST_LAST_MEMBER_CNT,
1205 br->multicast_last_member_count) ||
1206 nla_put_u32(skb, IFLA_BR_MCAST_STARTUP_QUERY_CNT,
1207 br->multicast_startup_query_count))
1208 return -EMSGSIZE;
1209
1210 clockval = jiffies_to_clock_t(br->multicast_last_member_interval);
1211 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_LAST_MEMBER_INTVL, clockval,
1212 IFLA_BR_PAD))
1213 return -EMSGSIZE;
1214 clockval = jiffies_to_clock_t(br->multicast_membership_interval);
1215 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_MEMBERSHIP_INTVL, clockval,
1216 IFLA_BR_PAD))
1217 return -EMSGSIZE;
1218 clockval = jiffies_to_clock_t(br->multicast_querier_interval);
1219 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERIER_INTVL, clockval,
1220 IFLA_BR_PAD))
1221 return -EMSGSIZE;
1222 clockval = jiffies_to_clock_t(br->multicast_query_interval);
1223 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_INTVL, clockval,
1224 IFLA_BR_PAD))
1225 return -EMSGSIZE;
1226 clockval = jiffies_to_clock_t(br->multicast_query_response_interval);
1227 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_RESPONSE_INTVL, clockval,
1228 IFLA_BR_PAD))
1229 return -EMSGSIZE;
1230 clockval = jiffies_to_clock_t(br->multicast_startup_query_interval);
1231 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_STARTUP_QUERY_INTVL, clockval,
1232 IFLA_BR_PAD))
1233 return -EMSGSIZE;
1234 #endif
1235 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1236 if (nla_put_u8(skb, IFLA_BR_NF_CALL_IPTABLES,
1237 br->nf_call_iptables ? 1 : 0) ||
1238 nla_put_u8(skb, IFLA_BR_NF_CALL_IP6TABLES,
1239 br->nf_call_ip6tables ? 1 : 0) ||
1240 nla_put_u8(skb, IFLA_BR_NF_CALL_ARPTABLES,
1241 br->nf_call_arptables ? 1 : 0))
1242 return -EMSGSIZE;
1243 #endif
1244
1245 return 0;
1246 }
1247
1248 static size_t bridge_get_linkxstats_size(const struct net_device *dev)
1249 {
1250 struct net_bridge *br = netdev_priv(dev);
1251 struct net_bridge_vlan_group *vg;
1252 struct net_bridge_vlan *v;
1253 int numvls = 0;
1254
1255 vg = br_vlan_group(br);
1256 if (vg) {
1257 /* we need to count all, even placeholder entries */
1258 list_for_each_entry(v, &vg->vlan_list, vlist)
1259 numvls++;
1260 }
1261
1262 return numvls * nla_total_size(sizeof(struct bridge_vlan_xstats)) +
1263 nla_total_size(sizeof(struct br_mcast_stats)) +
1264 nla_total_size(0);
1265 }
1266
1267 static size_t brport_get_linkxstats_size(const struct net_device *dev)
1268 {
1269 return nla_total_size(sizeof(struct br_mcast_stats)) +
1270 nla_total_size(0);
1271 }
1272
1273 static size_t br_get_linkxstats_size(const struct net_device *dev, int attr)
1274 {
1275 size_t retsize = 0;
1276
1277 switch (attr) {
1278 case IFLA_STATS_LINK_XSTATS:
1279 retsize = bridge_get_linkxstats_size(dev);
1280 break;
1281 case IFLA_STATS_LINK_XSTATS_SLAVE:
1282 retsize = brport_get_linkxstats_size(dev);
1283 break;
1284 }
1285
1286 return retsize;
1287 }
1288
1289 static int bridge_fill_linkxstats(struct sk_buff *skb,
1290 const struct net_device *dev,
1291 int *prividx)
1292 {
1293 struct net_bridge *br = netdev_priv(dev);
1294 struct nlattr *nla __maybe_unused;
1295 struct net_bridge_vlan_group *vg;
1296 struct net_bridge_vlan *v;
1297 struct nlattr *nest;
1298 int vl_idx = 0;
1299
1300 nest = nla_nest_start(skb, LINK_XSTATS_TYPE_BRIDGE);
1301 if (!nest)
1302 return -EMSGSIZE;
1303
1304 vg = br_vlan_group(br);
1305 if (vg) {
1306 list_for_each_entry(v, &vg->vlan_list, vlist) {
1307 struct bridge_vlan_xstats vxi;
1308 struct br_vlan_stats stats;
1309
1310 if (++vl_idx < *prividx)
1311 continue;
1312 memset(&vxi, 0, sizeof(vxi));
1313 vxi.vid = v->vid;
1314 br_vlan_get_stats(v, &stats);
1315 vxi.rx_bytes = stats.rx_bytes;
1316 vxi.rx_packets = stats.rx_packets;
1317 vxi.tx_bytes = stats.tx_bytes;
1318 vxi.tx_packets = stats.tx_packets;
1319
1320 if (nla_put(skb, BRIDGE_XSTATS_VLAN, sizeof(vxi), &vxi))
1321 goto nla_put_failure;
1322 }
1323 }
1324
1325 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1326 if (++vl_idx >= *prividx) {
1327 nla = nla_reserve_64bit(skb, BRIDGE_XSTATS_MCAST,
1328 sizeof(struct br_mcast_stats),
1329 BRIDGE_XSTATS_PAD);
1330 if (!nla)
1331 goto nla_put_failure;
1332 br_multicast_get_stats(br, NULL, nla_data(nla));
1333 }
1334 #endif
1335 nla_nest_end(skb, nest);
1336 *prividx = 0;
1337
1338 return 0;
1339
1340 nla_put_failure:
1341 nla_nest_end(skb, nest);
1342 *prividx = vl_idx;
1343
1344 return -EMSGSIZE;
1345 }
1346
1347 static int brport_fill_linkxstats(struct sk_buff *skb,
1348 const struct net_device *dev,
1349 int *prividx)
1350 {
1351 struct net_bridge_port *p = br_port_get_rtnl(dev);
1352 struct nlattr *nla __maybe_unused;
1353 struct nlattr *nest;
1354
1355 if (!p)
1356 return 0;
1357
1358 nest = nla_nest_start(skb, LINK_XSTATS_TYPE_BRIDGE);
1359 if (!nest)
1360 return -EMSGSIZE;
1361 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1362 nla = nla_reserve_64bit(skb, BRIDGE_XSTATS_MCAST,
1363 sizeof(struct br_mcast_stats),
1364 BRIDGE_XSTATS_PAD);
1365 if (!nla) {
1366 nla_nest_end(skb, nest);
1367 return -EMSGSIZE;
1368 }
1369 br_multicast_get_stats(p->br, p, nla_data(nla));
1370 #endif
1371 nla_nest_end(skb, nest);
1372
1373 return 0;
1374 }
1375
1376 static int br_fill_linkxstats(struct sk_buff *skb, const struct net_device *dev,
1377 int *prividx, int attr)
1378 {
1379 int ret = -EINVAL;
1380
1381 switch (attr) {
1382 case IFLA_STATS_LINK_XSTATS:
1383 ret = bridge_fill_linkxstats(skb, dev, prividx);
1384 break;
1385 case IFLA_STATS_LINK_XSTATS_SLAVE:
1386 ret = brport_fill_linkxstats(skb, dev, prividx);
1387 break;
1388 }
1389
1390 return ret;
1391 }
1392
1393 static struct rtnl_af_ops br_af_ops __read_mostly = {
1394 .family = AF_BRIDGE,
1395 .get_link_af_size = br_get_link_af_size_filtered,
1396 };
1397
1398 struct rtnl_link_ops br_link_ops __read_mostly = {
1399 .kind = "bridge",
1400 .priv_size = sizeof(struct net_bridge),
1401 .setup = br_dev_setup,
1402 .maxtype = IFLA_BR_MAX,
1403 .policy = br_policy,
1404 .validate = br_validate,
1405 .newlink = br_dev_newlink,
1406 .changelink = br_changelink,
1407 .dellink = br_dev_delete,
1408 .get_size = br_get_size,
1409 .fill_info = br_fill_info,
1410 .fill_linkxstats = br_fill_linkxstats,
1411 .get_linkxstats_size = br_get_linkxstats_size,
1412
1413 .slave_maxtype = IFLA_BRPORT_MAX,
1414 .slave_policy = br_port_policy,
1415 .slave_changelink = br_port_slave_changelink,
1416 .get_slave_size = br_port_get_slave_size,
1417 .fill_slave_info = br_port_fill_slave_info,
1418 };
1419
1420 int __init br_netlink_init(void)
1421 {
1422 int err;
1423
1424 br_mdb_init();
1425 rtnl_af_register(&br_af_ops);
1426
1427 err = rtnl_link_register(&br_link_ops);
1428 if (err)
1429 goto out_af;
1430
1431 return 0;
1432
1433 out_af:
1434 rtnl_af_unregister(&br_af_ops);
1435 br_mdb_uninit();
1436 return err;
1437 }
1438
1439 void br_netlink_fini(void)
1440 {
1441 br_mdb_uninit();
1442 rtnl_af_unregister(&br_af_ops);
1443 rtnl_link_unregister(&br_link_ops);
1444 }
This page took 0.068965 seconds and 5 git commands to generate.