/linux/net/netfilter/ |
A D | nf_flow_table_offload.c | 732 flow_rule = kzalloc(sizeof(*flow_rule), GFP_KERNEL); in nf_flow_offload_rule_alloc() 733 if (!flow_rule) in nf_flow_offload_rule_alloc() 737 if (!flow_rule->rule) in nf_flow_offload_rule_alloc() 740 flow_rule->rule->match.dissector = &flow_rule->match.dissector; in nf_flow_offload_rule_alloc() 741 flow_rule->rule->match.mask = &flow_rule->match.mask; in nf_flow_offload_rule_alloc() 742 flow_rule->rule->match.key = &flow_rule->match.key; in nf_flow_offload_rule_alloc() 757 return flow_rule; in nf_flow_offload_rule_alloc() 762 kfree(flow_rule); in nf_flow_offload_rule_alloc() 780 kfree(flow_rule); in __nf_flow_offload_destroy() 798 if (!flow_rule[0]) in nf_flow_offload_alloc() [all …]
|
A D | nf_flow_table_inet.c | 27 struct nf_flow_rule *flow_rule) in nf_flow_rule_route_inet() argument 34 err = nf_flow_rule_route_ipv4(net, flow, dir, flow_rule); in nf_flow_rule_route_inet() 37 err = nf_flow_rule_route_ipv6(net, flow, dir, flow_rule); in nf_flow_rule_route_inet()
|
/linux/include/net/ |
A D | flow_offload.h | 75 struct flow_rule; 77 void flow_rule_match_meta(const struct flow_rule *rule, 79 void flow_rule_match_basic(const struct flow_rule *rule, 85 void flow_rule_match_vlan(const struct flow_rule *rule, 93 void flow_rule_match_ip(const struct flow_rule *rule, 97 void flow_rule_match_tcp(const struct flow_rule *rule, 99 void flow_rule_match_icmp(const struct flow_rule *rule, 117 void flow_rule_match_ct(const struct flow_rule *rule, 384 struct flow_rule { struct 550 struct flow_rule *rule; [all …]
|
/linux/net/core/ |
A D | flow_offload.c | 9 struct flow_rule *flow_rule_alloc(unsigned int num_actions) in flow_rule_alloc() 11 struct flow_rule *rule; in flow_rule_alloc() 37 void flow_rule_match_meta(const struct flow_rule *rule, in flow_rule_match_meta() 44 void flow_rule_match_basic(const struct flow_rule *rule, in flow_rule_match_basic() 65 void flow_rule_match_vlan(const struct flow_rule *rule, in flow_rule_match_vlan() 72 void flow_rule_match_cvlan(const struct flow_rule *rule, in flow_rule_match_cvlan() 93 void flow_rule_match_ip(const struct flow_rule *rule, in flow_rule_match_ip() 107 void flow_rule_match_tcp(const struct flow_rule *rule, in flow_rule_match_tcp() 114 void flow_rule_match_icmp(const struct flow_rule *rule, in flow_rule_match_icmp() 121 void flow_rule_match_mpls(const struct flow_rule *rule, in flow_rule_match_mpls() [all …]
|
/linux/drivers/net/ethernet/netronome/nfp/flower/ |
A D | main.h | 422 struct flow_rule *rule); 436 struct flow_rule *rule, 441 struct flow_rule *rule); 445 struct flow_rule *rule); 457 struct flow_rule *rule); 461 struct flow_rule *rule); 465 struct flow_rule *rule); 469 struct flow_rule *rule); 471 struct flow_rule *rule, 478 struct flow_rule *rule, [all …]
|
A D | match.c | 25 struct flow_rule *rule) in nfp_flower_compile_tci() 119 struct flow_rule *rule, in nfp_flower_compile_mpls() 174 struct flow_rule *rule, in nfp_flower_compile_mac_mpls() 188 struct flow_rule *rule) in nfp_flower_compile_tport() 308 struct flow_rule *rule) in nfp_flower_compile_vlan() 380 struct flow_rule *rule) in nfp_flower_compile_tun_ipv4_addrs() 396 struct flow_rule *rule) in nfp_flower_compile_tun_ipv6_addrs() 466 struct flow_rule *rule) in nfp_flower_compile_ipv4_gre_tun() 481 struct flow_rule *rule) in nfp_flower_compile_ipv4_udp_tun() 491 struct flow_rule *rule) in nfp_flower_compile_ipv6_udp_tun() [all …]
|
A D | conntrack.c | 23 static struct flow_action_entry *get_flow_act(struct flow_rule *rule, 67 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in is_post_ct_flow() 256 struct flow_rule *rule) in nfp_ct_check_mangle_merge() 475 static int nfp_fl_merge_actions_offload(struct flow_rule **rules, in nfp_fl_merge_actions_offload() 482 struct flow_rule *a_rule; in nfp_fl_merge_actions_offload() 551 struct flow_rule *rules[_CT_TYPE_MAX]; in nfp_fl_ct_add_offload() 1353 static struct flow_action_entry *get_flow_act(struct flow_rule *rule, in get_flow_act() 1471 struct flow_rule *rule = flow_cls_offload_flow_rule(flow); in nfp_fl_ct_handle_post_ct()
|
A D | action.c | 265 nfp_flower_tun_is_gre(struct flow_rule *rule, int start_idx) in nfp_flower_tun_is_gre() 283 struct flow_rule *rule, in nfp_fl_get_tun_from_act() 792 nfp_fl_commit_mangle(struct flow_rule *rule, char *nfp_action, in nfp_fl_commit_mangle() 980 struct flow_rule *rule, in nfp_flower_loop_action() 1198 struct flow_rule *rule, in nfp_flower_compile_action()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | rdma.c | 26 struct mlx5_flow_handle *flow_rule = NULL; in mlx5_rdma_enable_roce_steering() local 90 flow_rule = mlx5_add_flow_rules(ft, spec, &flow_act, NULL, 0); in mlx5_rdma_enable_roce_steering() 91 if (IS_ERR(flow_rule)) { in mlx5_rdma_enable_roce_steering() 92 err = PTR_ERR(flow_rule); in mlx5_rdma_enable_roce_steering() 102 roce->allow_rule = flow_rule; in mlx5_rdma_enable_roce_steering()
|
A D | en_rep.c | 331 struct mlx5_flow_handle *flow_rule; in mlx5e_sqs2vport_start() local 355 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start() 356 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start() 366 if (IS_ERR(flow_rule)) { in mlx5e_sqs2vport_start() 367 err = PTR_ERR(flow_rule); in mlx5e_sqs2vport_start() 763 struct mlx5_flow_handle *flow_rule; in mlx5e_create_rep_vport_rx_rule() local 770 if (IS_ERR(flow_rule)) in mlx5e_create_rep_vport_rx_rule() 771 return PTR_ERR(flow_rule); in mlx5e_create_rep_vport_rx_rule() 772 rpriv->vport_rx_rule = flow_rule; in mlx5e_create_rep_vport_rx_rule() 1329 if (IS_ERR(flow_rule)) in mlx5e_vport_rep_event_pair() [all …]
|
A D | eswitch.c | 61 struct mlx5_flow_handle *flow_rule; member 231 flow_rule = in __esw_fdb_set_vport_rule() 234 if (IS_ERR(flow_rule)) { in __esw_fdb_set_vport_rule() 238 flow_rule = NULL; in __esw_fdb_set_vport_rule() 242 return flow_rule; in __esw_fdb_set_vport_rule() 334 if (vaddr->flow_rule) in esw_del_uc_addr() 336 vaddr->flow_rule = NULL; in esw_del_uc_addr() 374 iter_vaddr->flow_rule = in update_allmulti_vports() 425 vport, mac, vaddr->flow_rule, in esw_add_mc_addr() 452 if (vaddr->flow_rule) in esw_del_mc_addr() [all …]
|
A D | eswitch_offloads.c | 1011 if (IS_ERR(flow_rule)) in mlx5_eswitch_add_send_to_vport_rule() 1013 PTR_ERR(flow_rule)); in mlx5_eswitch_add_send_to_vport_rule() 1016 return flow_rule; in mlx5_eswitch_add_send_to_vport_rule() 1082 if (IS_ERR(flow_rule)) { in mlx5_eswitch_add_send_to_vport_meta_rules() 1364 if (IS_ERR(flow_rule)) { in esw_add_fdb_miss_rule() 1379 if (IS_ERR(flow_rule)) { in esw_add_fdb_miss_rule() 1432 if (IS_ERR(flow_rule)) in esw_add_restore_rule() 1437 return flow_rule; in esw_add_restore_rule() 2011 if (IS_ERR(flow_rule)) { in mlx5_eswitch_create_vport_rx_rule() 2018 return flow_rule; in mlx5_eswitch_create_vport_rx_rule() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | tc_ct.c | 80 struct mlx5_flow_handle *flow_rule; member 262 struct flow_rule *rule) in mlx5_tc_ct_rule_to_tuple_nat() 324 struct flow_rule *rule) in mlx5_tc_ct_set_tuple_match() 474 mlx5_tc_ct_get_ct_metadata_action(struct flow_rule *flow_rule) in mlx5_tc_ct_get_ct_metadata_action() argument 606 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_nat() argument 651 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_create_mod_hdr() argument 731 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_add_rule() argument 964 struct flow_rule *flow_rule, in mlx5_tc_ct_entry_add_rules() argument 1003 struct flow_rule *flow_rule = flow_cls_offload_flow_rule(flow); in mlx5_tc_ct_block_flow_offload_add() local 1451 pre_ct->flow_rule = rule; in tc_ct_pre_ct_add_rules() [all …]
|
A D | tc_tun_vxlan.c | 21 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_vxlan() 107 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_vxlan()
|
A D | tc_tun_geneve.c | 25 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_check_udp_dport_geneve() 127 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_vni() 161 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_geneve_options()
|
A D | tc_tun_gre.c | 63 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlx5e_tc_tun_parse_gretap()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
A D | legacy.c | 254 struct mlx5_flow_handle *flow_rule; in _mlx5_eswitch_set_vepa_locked() local 282 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, spec, in _mlx5_eswitch_set_vepa_locked() 284 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked() 285 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked() 288 esw->fdb_table.legacy.vepa_uplink_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked() 296 flow_rule = mlx5_add_flow_rules(esw->fdb_table.legacy.vepa_fdb, NULL, in _mlx5_eswitch_set_vepa_locked() 298 if (IS_ERR(flow_rule)) { in _mlx5_eswitch_set_vepa_locked() 299 err = PTR_ERR(flow_rule); in _mlx5_eswitch_set_vepa_locked() 302 esw->fdb_table.legacy.vepa_star_rule = flow_rule; in _mlx5_eswitch_set_vepa_locked()
|
/linux/include/net/netfilter/ |
A D | nf_flow_table.h | 47 struct flow_rule *rule; 60 struct nf_flow_rule *flow_rule); 308 struct nf_flow_rule *flow_rule); 311 struct nf_flow_rule *flow_rule);
|
A D | nf_tables_offload.h | 67 struct flow_rule *rule;
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/tc/ |
A D | int_port.c | 68 struct mlx5_flow_handle *flow_rule; in mlx5e_int_port_create_rx_rule() local 93 flow_rule = mlx5_add_flow_rules(esw->offloads.ft_offloads, spec, in mlx5e_int_port_create_rx_rule() 95 if (IS_ERR(flow_rule)) in mlx5e_int_port_create_rx_rule() 97 PTR_ERR(flow_rule)); in mlx5e_int_port_create_rx_rule() 101 return flow_rule; in mlx5e_int_port_create_rx_rule()
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
A D | spectrum_flower.c | 243 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_meta() 335 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ports() 362 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_tcp() 393 const struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse_ip() 426 struct flow_rule *rule = flow_cls_offload_flow_rule(f); in mlxsw_sp_flower_parse()
|
/linux/drivers/net/ethernet/stmicro/stmmac/ |
A D | stmmac_tc.c | 458 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_basic_flow() 475 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ip4_flow() 510 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_ports_flow() 586 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_flow() 660 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in tc_add_vlan_flow()
|
/linux/drivers/net/ethernet/freescale/dpaa2/ |
A D | dpaa2-switch-flower.c | 14 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_key() 488 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace_acl() 532 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_flower_parse_mirror_key() 638 struct flow_rule *rule = flow_cls_offload_flow_rule(cls); in dpaa2_switch_cls_flower_replace()
|
/linux/drivers/net/ethernet/chelsio/cxgb4/ |
A D | cxgb4_tc_flower.h | 140 int cxgb4_flow_rule_replace(struct net_device *dev, struct flow_rule *rule,
|
/linux/drivers/net/ethernet/qlogic/qede/ |
A D | qede_filter.c | 1703 qede_flow_parse_ports(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_ports() 1724 qede_flow_parse_v6_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v6_common() 1756 qede_flow_parse_v4_common(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_v4_common() 1780 qede_flow_parse_tcp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v6() 1790 qede_flow_parse_tcp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_tcp_v4() 1800 qede_flow_parse_udp_v6(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v6() 1810 qede_flow_parse_udp_v4(struct qede_dev *edev, struct flow_rule *rule, in qede_flow_parse_udp_v4() 1821 struct flow_rule *rule, struct qede_arfs_tuple *tuple) in qede_parse_flow_attr()
|