Lines Matching refs:vlan

405 	struct net_bridge_vlan *vlan, *tmp;  in __vlan_flush()  local
409 list_for_each_entry_safe(vlan, tmp, &vg->vlan_list, vlist) { in __vlan_flush()
412 v_start = vlan->vid; in __vlan_flush()
413 } else if (vlan->vid - v_end != 1) { in __vlan_flush()
416 v_start = vlan->vid; in __vlan_flush()
418 v_end = vlan->vid; in __vlan_flush()
420 __vlan_del(vlan); in __vlan_flush()
493 struct net_bridge_vlan **vlan) in __allowed_ingress() argument
587 *vlan = v; in __allowed_ingress()
599 struct net_bridge_vlan **vlan) in br_allowed_ingress() argument
604 *vlan = NULL; in br_allowed_ingress()
610 return __allowed_ingress(br, vg, skb, vid, state, vlan); in br_allowed_ingress()
669 struct net_bridge_vlan *vlan, in br_vlan_add_existing() argument
675 err = br_switchdev_port_vlan_add(br->dev, vlan->vid, flags, extack); in br_vlan_add_existing()
679 if (!br_vlan_is_brentry(vlan)) { in br_vlan_add_existing()
686 err = br_fdb_add_local(br, NULL, br->dev->dev_addr, vlan->vid); in br_vlan_add_existing()
692 refcount_inc(&vlan->refcnt); in br_vlan_add_existing()
693 vlan->flags |= BRIDGE_VLAN_INFO_BRENTRY; in br_vlan_add_existing()
696 br_multicast_toggle_one_vlan(vlan, true); in br_vlan_add_existing()
699 if (__vlan_add_flags(vlan, flags)) in br_vlan_add_existing()
706 br_switchdev_port_vlan_del(br->dev, vlan->vid); in br_vlan_add_existing()
718 struct net_bridge_vlan *vlan; in br_vlan_add() local
725 vlan = br_vlan_find(vg, vid); in br_vlan_add()
726 if (vlan) in br_vlan_add()
727 return br_vlan_add_existing(br, vg, vlan, flags, changed, in br_vlan_add()
730 vlan = kzalloc(sizeof(*vlan), GFP_KERNEL); in br_vlan_add()
731 if (!vlan) in br_vlan_add()
734 vlan->stats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats); in br_vlan_add()
735 if (!vlan->stats) { in br_vlan_add()
736 kfree(vlan); in br_vlan_add()
739 vlan->vid = vid; in br_vlan_add()
740 vlan->flags = flags | BRIDGE_VLAN_INFO_MASTER; in br_vlan_add()
741 vlan->flags &= ~BRIDGE_VLAN_INFO_PVID; in br_vlan_add()
742 vlan->br = br; in br_vlan_add()
744 refcount_set(&vlan->refcnt, 1); in br_vlan_add()
745 ret = __vlan_add(vlan, flags, extack); in br_vlan_add()
747 free_percpu(vlan->stats); in br_vlan_add()
748 kfree(vlan); in br_vlan_add()
891 struct net_bridge_vlan *vlan; in __br_vlan_set_proto() local
905 list_for_each_entry(vlan, &vg->vlan_list, vlist) { in __br_vlan_set_proto()
906 err = vlan_vid_add(p->dev, proto, vlan->vid); in __br_vlan_set_proto()
920 list_for_each_entry(vlan, &vg->vlan_list, vlist) in __br_vlan_set_proto()
921 vlan_vid_del(p->dev, oldproto, vlan->vid); in __br_vlan_set_proto()
930 list_for_each_entry_continue_reverse(vlan, &vg->vlan_list, vlist) in __br_vlan_set_proto()
931 vlan_vid_del(p->dev, proto, vlan->vid); in __br_vlan_set_proto()
935 list_for_each_entry(vlan, &vg->vlan_list, vlist) in __br_vlan_set_proto()
936 vlan_vid_del(p->dev, proto, vlan->vid); in __br_vlan_set_proto()
1242 struct net_bridge_vlan *vlan; in nbp_vlan_add() local
1248 vlan = br_vlan_find(nbp_vlan_group(port), vid); in nbp_vlan_add()
1249 if (vlan) { in nbp_vlan_add()
1254 *changed = __vlan_add_flags(vlan, flags); in nbp_vlan_add()
1259 vlan = kzalloc(sizeof(*vlan), GFP_KERNEL); in nbp_vlan_add()
1260 if (!vlan) in nbp_vlan_add()
1263 vlan->vid = vid; in nbp_vlan_add()
1264 vlan->port = port; in nbp_vlan_add()
1265 ret = __vlan_add(vlan, flags, extack); in nbp_vlan_add()
1267 kfree(vlan); in nbp_vlan_add()
1385 ctx->vlan[idx].proto == br->vlan_proto) { in br_vlan_fill_forward_path_pvid()
1386 vid = ctx->vlan[idx].id; in br_vlan_fill_forward_path_pvid()
1576 struct net_bridge_vlan *vlan; in br_vlan_set_all_vlan_dev_state() local
1579 list_for_each_entry(vlan, &vg->vlan_list, vlist) { in br_vlan_set_all_vlan_dev_state()
1581 vlan->vid); in br_vlan_set_all_vlan_dev_state()