/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/esw/acl/ |
H A D | ingress_ofld.c | 43 if (vport->ingress.offloads.modify_metadata_rule) { in esw_acl_ingress_prio_tag_create() 45 flow_act.modify_hdr = vport->ingress.offloads.modify_metadata; in esw_acl_ingress_prio_tag_create() 82 vport->ingress.offloads.modify_metadata = in esw_acl_ingress_mod_metadata_create() 85 if (IS_ERR(vport->ingress.offloads.modify_metadata)) { in esw_acl_ingress_mod_metadata_create() 86 err = PTR_ERR(vport->ingress.offloads.modify_metadata); in esw_acl_ingress_mod_metadata_create() 94 flow_act.modify_hdr = vport->ingress.offloads.modify_metadata; in esw_acl_ingress_mod_metadata_create() 95 flow_act.fg = vport->ingress.offloads.metadata_allmatch_grp; in esw_acl_ingress_mod_metadata_create() 96 vport->ingress.offloads.modify_metadata_rule = in esw_acl_ingress_mod_metadata_create() 99 if (IS_ERR(vport->ingress.offloads.modify_metadata_rule)) { in esw_acl_ingress_mod_metadata_create() 100 err = PTR_ERR(vport->ingress.offloads in esw_acl_ingress_mod_metadata_create() [all...] |
H A D | egress_ofld.c | 11 if (!vport->egress.offloads.fwd_rule) in esw_acl_egress_ofld_fwd2vport_destroy() 14 mlx5_del_flow_rules(vport->egress.offloads.fwd_rule); in esw_acl_egress_ofld_fwd2vport_destroy() 15 vport->egress.offloads.fwd_rule = NULL; in esw_acl_egress_ofld_fwd2vport_destroy() 21 xa_load(&vport->egress.offloads.bounce_rules, rule_index); in esw_acl_egress_ofld_bounce_rule_destroy() 27 xa_erase(&vport->egress.offloads.bounce_rules, rule_index); in esw_acl_egress_ofld_bounce_rule_destroy() 35 xa_for_each(&vport->egress.offloads.bounce_rules, i, bounce_rule) { in esw_acl_egress_ofld_bounce_rules_destroy() 37 xa_erase(&vport->egress.offloads.bounce_rules, i); in esw_acl_egress_ofld_bounce_rules_destroy() 56 vport->egress.offloads.fwd_rule = in esw_acl_egress_ofld_fwd2vport_create() 59 if (IS_ERR(vport->egress.offloads.fwd_rule)) { in esw_acl_egress_ofld_fwd2vport_create() 60 err = PTR_ERR(vport->egress.offloads in esw_acl_egress_ofld_fwd2vport_create() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/esw/acl/ |
H A D | ingress_ofld.c | 43 if (vport->ingress.offloads.modify_metadata_rule) { in esw_acl_ingress_prio_tag_create() 45 flow_act.modify_hdr = vport->ingress.offloads.modify_metadata; in esw_acl_ingress_prio_tag_create() 82 vport->ingress.offloads.modify_metadata = in esw_acl_ingress_mod_metadata_create() 85 if (IS_ERR(vport->ingress.offloads.modify_metadata)) { in esw_acl_ingress_mod_metadata_create() 86 err = PTR_ERR(vport->ingress.offloads.modify_metadata); in esw_acl_ingress_mod_metadata_create() 94 flow_act.modify_hdr = vport->ingress.offloads.modify_metadata; in esw_acl_ingress_mod_metadata_create() 95 vport->ingress.offloads.modify_metadata_rule = in esw_acl_ingress_mod_metadata_create() 98 if (IS_ERR(vport->ingress.offloads.modify_metadata_rule)) { in esw_acl_ingress_mod_metadata_create() 99 err = PTR_ERR(vport->ingress.offloads.modify_metadata_rule); in esw_acl_ingress_mod_metadata_create() 103 mlx5_modify_header_dealloc(esw->dev, vport->ingress.offloads in esw_acl_ingress_mod_metadata_create() [all...] |
H A D | egress_ofld.c | 11 if (!vport->egress.offloads.fwd_rule) in esw_acl_egress_ofld_fwd2vport_destroy() 14 mlx5_del_flow_rules(vport->egress.offloads.fwd_rule); in esw_acl_egress_ofld_fwd2vport_destroy() 15 vport->egress.offloads.fwd_rule = NULL; in esw_acl_egress_ofld_fwd2vport_destroy() 33 vport->egress.offloads.fwd_rule = in esw_acl_egress_ofld_fwd2vport_create() 36 if (IS_ERR(vport->egress.offloads.fwd_rule)) { in esw_acl_egress_ofld_fwd2vport_create() 37 err = PTR_ERR(vport->egress.offloads.fwd_rule); in esw_acl_egress_ofld_fwd2vport_create() 41 vport->egress.offloads.fwd_rule = NULL; in esw_acl_egress_ofld_fwd2vport_create() 132 vport->egress.offloads.fwd_grp = fwd_grp; in esw_acl_egress_ofld_groups_create() 144 if (!IS_ERR_OR_NULL(vport->egress.offloads.fwd_grp)) { in esw_acl_egress_ofld_groups_destroy() 145 mlx5_destroy_flow_group(vport->egress.offloads in esw_acl_egress_ofld_groups_destroy() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eswitch_offloads.c | 120 hash_for_each_possible(esw->fdb_table.offloads.vports.table, e, hlist, key) in esw_vport_tbl_lookup() 134 mutex_lock(&esw->fdb_table.offloads.vports.lock); in esw_vport_tbl_put() 144 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in esw_vport_tbl_put() 157 mutex_lock(&esw->fdb_table.offloads.vports.lock); in esw_vport_tbl_get() 185 hash_add(esw->fdb_table.offloads.vports.table, &e->hlist, hkey); in esw_vport_tbl_get() 187 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in esw_vport_tbl_get() 193 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in esw_vport_tbl_get() 241 return &esw->offloads.vport_reps[idx]; in mlx5_eswitch_get_rep() 427 atomic64_inc(&esw->offloads.num_flows); in mlx5_eswitch_add_offloaded_rule() 501 atomic64_inc(&esw->offloads in mlx5_eswitch_add_fwd_rule() 645 struct offloads_fdb *offloads = &esw->fdb_table.offloads; mlx5_eswitch_add_vlan_action() local 714 struct offloads_fdb *offloads = &esw->fdb_table.offloads; mlx5_eswitch_del_vlan_action() local 1504 struct mlx5_esw_offload *offloads = &esw->offloads; esw_destroy_offloads_table() local 1635 struct mlx5_esw_offload *offloads = &esw->offloads; esw_destroy_restore_table() local [all...] |
H A D | eswitch_offloads_termtbl.c | 120 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 122 hash_for_each_possible(esw->offloads.termtbl_tbl, tt, in mlx5_eswitch_termtbl_get_create() 150 hash_add(esw->offloads.termtbl_tbl, &tt->termtbl_hlist, hash_key); in mlx5_eswitch_termtbl_get_create() 153 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 157 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 165 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put() 168 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put()
|
H A D | eswitch.h | 67 ((esw)->fdb_table.offloads.esw_chains_priv) 91 } offloads; member 107 } offloads; member 192 } offloads; member 272 struct mlx5_esw_offload offloads; member 590 (rep) = &(esw)->offloads.vport_reps[i], \ 595 (rep) = &(esw)->offloads.vport_reps[i], \ 600 (rep) = &(esw)->offloads.vport_reps[i], \ 612 (rep) = &(esw)->offloads.vport_reps[i], \ 617 (rep) = &(esw)->offloads [all...] |
H A D | en_tc.c | 445 &esw->offloads.mod_hdr : in get_mod_hdr_table() 1731 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_tc_update_neigh_used_value() 1748 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_tc_update_neigh_used_value() 1806 if (!refcount_dec_and_mutex_lock(&e->refcnt, &esw->offloads.encap_tbl_lock)) in mlx5e_encap_put() 1809 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_encap_put() 1818 if (!refcount_dec_and_mutex_lock(&d->refcnt, &esw->offloads.decap_tbl_lock)) in mlx5e_decap_put() 1821 mutex_unlock(&esw->offloads.decap_tbl_lock); in mlx5e_decap_put() 1836 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_detach_encap() 1840 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_detach_encap() 1844 mutex_unlock(&esw->offloads in mlx5e_detach_encap() [all...] |
H A D | eswitch.c | 1591 * mlx5_eswitch_enable_locked() Enables eswitch in either legacy or offloads 1774 mutex_init(&esw->offloads.encap_tbl_lock); in mlx5_eswitch_init() 1775 hash_init(esw->offloads.encap_tbl); in mlx5_eswitch_init() 1776 mutex_init(&esw->offloads.decap_tbl_lock); in mlx5_eswitch_init() 1777 hash_init(esw->offloads.decap_tbl); in mlx5_eswitch_init() 1778 mlx5e_mod_hdr_tbl_init(&esw->offloads.mod_hdr); in mlx5_eswitch_init() 1779 atomic64_set(&esw->offloads.num_flows, 0); in mlx5_eswitch_init() 1780 ida_init(&esw->offloads.vport_metadata_ida); in mlx5_eswitch_init() 1794 esw->offloads.inline_mode = MLX5_INLINE_MODE_NONE; in mlx5_eswitch_init() 1818 ida_destroy(&esw->offloads in mlx5_eswitch_cleanup() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
H A D | vporttbl.c | 27 if (esw->offloads.encap != DEVLINK_ESWITCH_ENCAP_MODE_NONE) in esw_vport_tbl_init() 73 hash_for_each_possible(esw->fdb_table.offloads.vports.table, e, hlist, key) in esw_vport_tbl_lookup() 90 mutex_lock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get() 119 hash_add(esw->fdb_table.offloads.vports.table, &e->hlist, hkey); in mlx5_esw_vporttbl_get() 121 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get() 127 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_get() 138 mutex_lock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_put() 149 mutex_unlock(&esw->fdb_table.offloads.vports.lock); in mlx5_esw_vporttbl_put()
|
H A D | indir_table.c | 293 hash_add(esw->fdb_table.offloads.indir->table, &e->hlist, in mlx5_esw_indir_table_entry_create() 316 hash_for_each_possible(esw->fdb_table.offloads.indir->table, e, hlist, key) in mlx5_esw_indir_table_entry_lookup() 330 mutex_lock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_get() 348 mutex_unlock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_get() 352 mutex_unlock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_get() 361 mutex_lock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_put() 381 mutex_unlock(&esw->fdb_table.offloads.indir->lock); in mlx5_esw_indir_table_put()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eswitch_offloads.c | 56 xa_for_each(&((esw)->offloads.vport_reps), i, rep) 78 return xa_load(&esw->offloads.vport_reps, vport_num); in mlx5_eswitch_get_rep() 433 if (esw->offloads.ft_ipsec_tx_pol && in esw_setup_uplink_fwd_ipsec_needed() 450 if (!esw->offloads.ft_ipsec_tx_pol) in esw_flow_dests_fwd_ipsec_check() 495 dest[dest_idx].ft = esw->offloads.ft_ipsec_tx_pol; in esw_setup_dest_fwd_ipsec() 769 atomic64_inc(&esw->offloads.num_flows); in mlx5_eswitch_add_offloaded_rule() 855 atomic64_inc(&esw->offloads.num_flows); in mlx5_eswitch_add_fwd_rule() 890 atomic64_dec(&esw->offloads.num_flows); in __mlx5_eswitch_del_rule() 992 on_esw == from_esw && on_esw->offloads.ft_ipsec_tx_pol) { in mlx5_eswitch_add_send_to_vport_rule() 993 dest.ft = on_esw->offloads in mlx5_eswitch_add_send_to_vport_rule() 2012 struct mlx5_esw_offload *offloads = &esw->offloads; esw_destroy_offloads_table() local 2223 struct mlx5_esw_offload *offloads = &esw->offloads; esw_destroy_restore_table() local [all...] |
H A D | eswitch_offloads_termtbl.c | 122 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 124 hash_for_each_possible(esw->offloads.termtbl_tbl, tt, in mlx5_eswitch_termtbl_get_create() 151 hash_add(esw->offloads.termtbl_tbl, &tt->termtbl_hlist, hash_key); in mlx5_eswitch_termtbl_get_create() 154 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 158 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_get_create() 166 mutex_lock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put() 169 mutex_unlock(&esw->offloads.termtbl_mutex); in mlx5_eswitch_termtbl_put()
|
H A D | eswitch.c | 1458 * mlx5_eswitch_enable_locked() Enables eswitch in either legacy or offloads 1875 mutex_init(&esw->offloads.encap_tbl_lock); in mlx5_eswitch_init() 1876 hash_init(esw->offloads.encap_tbl); in mlx5_eswitch_init() 1877 mutex_init(&esw->offloads.decap_tbl_lock); in mlx5_eswitch_init() 1878 hash_init(esw->offloads.decap_tbl); in mlx5_eswitch_init() 1879 mlx5e_mod_hdr_tbl_init(&esw->offloads.mod_hdr); in mlx5_eswitch_init() 1880 atomic64_set(&esw->offloads.num_flows, 0); in mlx5_eswitch_init() 1881 ida_init(&esw->offloads.vport_metadata_ida); in mlx5_eswitch_init() 1882 xa_init_flags(&esw->offloads.vhca_map, XA_FLAGS_ALLOC); in mlx5_eswitch_init() 1889 esw->offloads in mlx5_eswitch_init() [all...] |
H A D | eswitch.h | 87 ((esw)->fdb_table.offloads.esw_chains_priv) 123 } offloads; member 147 } offloads; member 267 } offloads; member 373 struct mlx5_esw_offload offloads; member 869 return esw->fdb_table.offloads.slow_fdb; in mlx5_eswitch_get_slow_fdb()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | tc_tun_encap.c | 412 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_tc_update_neigh_used_value() 429 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_tc_update_neigh_used_value() 487 if (!refcount_dec_and_mutex_lock(&e->refcnt, &esw->offloads.encap_tbl_lock)) in mlx5e_encap_put() 491 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_encap_put() 500 lockdep_assert_held(&esw->offloads.encap_tbl_lock); in mlx5e_encap_put_locked() 513 if (!refcount_dec_and_mutex_lock(&d->refcnt, &esw->offloads.decap_tbl_lock)) in mlx5e_decap_put() 516 mutex_unlock(&esw->offloads.decap_tbl_lock); in mlx5e_decap_put() 544 mutex_lock(&esw->offloads.encap_tbl_lock); in mlx5e_detach_encap() 548 mutex_unlock(&esw->offloads.encap_tbl_lock); in mlx5e_detach_encap() 553 mutex_unlock(&esw->offloads in mlx5e_detach_encap() [all...] |
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | offload.c | 113 list_add_tail(&offload->offloads, &ondev->progs); in bpf_prog_offload_init() 222 list_del_init(&offload->offloads); in __bpf_prog_offload_destroy() 398 list_add_tail(&offmap->offloads, &ondev->maps); in bpf_map_offload_map_alloc() 416 list_del_init(&offmap->offloads); in __bpf_map_offload_destroy() 650 list_for_each_entry(offload, &ondev->progs, offloads) in bpf_offload_dev_netdev_unregister() 654 list_for_each_entry(offmap, &ondev->maps, offloads) in bpf_offload_dev_netdev_unregister() 658 list_for_each_entry_safe(offload, ptmp, &ondev->progs, offloads) in bpf_offload_dev_netdev_unregister() 660 list_for_each_entry_safe(offmap, mtmp, &ondev->maps, offloads) in bpf_offload_dev_netdev_unregister()
|
/kernel/linux/linux-5.10/drivers/net/hyperv/ |
H A D | rndis_filter.c | 1252 struct ndis_offload_params offloads; in rndis_netdev_set_hwcaps() local 1262 memset(&offloads, 0, sizeof(struct ndis_offload_params)); in rndis_netdev_set_hwcaps() 1265 offloads.ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_DISABLED; in rndis_netdev_set_hwcaps() 1281 offloads.tcp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1284 offloads.lso_v2_ipv4 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED; in rndis_netdev_set_hwcaps() 1292 offloads.udp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1300 offloads.tcp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1305 offloads.lso_v2_ipv6 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED; in rndis_netdev_set_hwcaps() 1313 offloads.udp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1322 offloads in rndis_netdev_set_hwcaps() [all...] |
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | offload.c | 115 list_del_init(&offload->offloads); in __bpf_prog_offload_destroy() 141 list_del_init(&offmap->offloads); in __bpf_map_offload_destroy() 169 list_for_each_entry(offload, &ondev->progs, offloads) in __bpf_offload_dev_netdev_unregister() 173 list_for_each_entry(offmap, &ondev->maps, offloads) in __bpf_offload_dev_netdev_unregister() 177 list_for_each_entry_safe(offload, ptmp, &ondev->progs, offloads) in __bpf_offload_dev_netdev_unregister() 179 list_for_each_entry_safe(offmap, mtmp, &ondev->maps, offloads) in __bpf_offload_dev_netdev_unregister() 220 list_add_tail(&offload->offloads, &ondev->progs); in __bpf_prog_dev_bound_init() 382 list_del_init(&prog->aux->offload->offloads); in bpf_prog_dev_bound_destroy() 542 list_add_tail(&offmap->offloads, &ondev->maps); in bpf_map_offload_map_alloc()
|
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | fou.c | 234 const struct net_offload **offloads; in fou_gro_receive() local 250 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in fou_gro_receive() 251 ops = rcu_dereference(offloads[proto]); in fou_gro_receive() 269 const struct net_offload **offloads; in fou_gro_complete() local 272 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in fou_gro_complete() 273 ops = rcu_dereference(offloads[proto]); in fou_gro_complete() 314 const struct net_offload **offloads; in gue_gro_receive() local 442 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in gue_gro_receive() 443 ops = rcu_dereference(offloads[proto]); in gue_gro_receive() 460 const struct net_offload **offloads; in gue_gro_complete() local [all...] |
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | fou_core.c | 235 const struct net_offload __rcu **offloads; in fou_gro_receive() local 251 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in fou_gro_receive() 252 ops = rcu_dereference(offloads[proto]); in fou_gro_receive() 265 const struct net_offload __rcu **offloads; in fou_gro_complete() local 270 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in fou_gro_complete() 271 ops = rcu_dereference(offloads[proto]); in fou_gro_complete() 310 const struct net_offload __rcu **offloads; in gue_gro_receive() local 434 offloads = NAPI_GRO_CB(skb)->is_ipv6 ? inet6_offloads : inet_offloads; in gue_gro_receive() 435 ops = rcu_dereference(offloads[proto]); in gue_gro_receive() 451 const struct net_offload __rcu **offloads; in gue_gro_complete() local [all...] |
/kernel/linux/linux-6.6/drivers/net/hyperv/ |
H A D | rndis_filter.c | 1354 struct ndis_offload_params offloads; in rndis_netdev_set_hwcaps() local 1364 memset(&offloads, 0, sizeof(struct ndis_offload_params)); in rndis_netdev_set_hwcaps() 1367 offloads.ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_DISABLED; in rndis_netdev_set_hwcaps() 1383 offloads.tcp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1386 offloads.lso_v2_ipv4 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED; in rndis_netdev_set_hwcaps() 1394 offloads.udp_ip_v4_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1402 offloads.tcp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1407 offloads.lso_v2_ipv6 = NDIS_OFFLOAD_PARAMETERS_LSOV2_ENABLED; in rndis_netdev_set_hwcaps() 1415 offloads.udp_ip_v6_csum = NDIS_OFFLOAD_PARAMETERS_TX_RX_ENABLED; in rndis_netdev_set_hwcaps() 1424 offloads in rndis_netdev_set_hwcaps() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_virtchnl.c | 380 /* VLAN offloads based on current device configuration */ in ice_vc_get_vlan_caps() 385 * and offloads being based on the inner VLAN or the in ice_vc_get_vlan_caps() 743 * RSS offloads 2137 * ice_vf_vlan_offload_ena - determine if capabilities support VLAN offloads 2731 supported_caps = &caps->offloads.stripping_support; in ice_vc_set_dvm_caps() 2737 supported_caps = &caps->offloads.insertion_support; in ice_vc_set_dvm_caps() 2743 caps->offloads.ethertype_init = VIRTCHNL_VLAN_ETHERTYPE_8100; in ice_vc_set_dvm_caps() 2744 caps->offloads.ethertype_match = in ice_vc_set_dvm_caps() 2757 supported_caps = &caps->offloads.stripping_support; in ice_vc_set_dvm_caps() 2768 supported_caps = &caps->offloads in ice_vc_set_dvm_caps() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/tc/ |
H A D | int_port.c | 93 flow_rule = mlx5_add_flow_rules(esw->offloads.ft_offloads, spec, in mlx5e_int_port_create_rx_rule() 96 mlx5_core_warn(esw->dev, "ft offloads: Failed to add internal vport rx rule err %ld\n", in mlx5e_int_port_create_rx_rule() 182 ctx = esw->offloads.reg_c0_obj_pool; in mlx5e_int_port_add() 233 ctx = esw->offloads.reg_c0_obj_pool; in mlx5e_int_port_remove()
|
/kernel/linux/linux-5.10/drivers/net/ |
H A D | virtio_net.c | 178 __virtio64 offloads; member 2442 static int virtnet_set_guest_offloads(struct virtnet_info *vi, u64 offloads) in virtnet_set_guest_offloads() argument 2445 vi->ctrl->offloads = cpu_to_virtio64(vi->vdev, offloads); in virtnet_set_guest_offloads() 2447 sg_init_one(&sg, &vi->ctrl->offloads, sizeof(vi->ctrl->offloads)); in virtnet_set_guest_offloads() 2460 u64 offloads = 0; in virtnet_clear_guest_offloads() local 2465 return virtnet_set_guest_offloads(vi, offloads); in virtnet_clear_guest_offloads() 2470 u64 offloads = vi->guest_offloads; in virtnet_restore_guest_offloads() local 2475 return virtnet_set_guest_offloads(vi, offloads); in virtnet_restore_guest_offloads() 2621 u64 offloads; virtnet_set_features() local [all...] |