/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_flow.c | 221 * @segs: array of one or more packet segments that describe the flow 225 ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) in ice_flow_val_hdrs() argument 231 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && in ice_flow_val_hdrs() 232 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) in ice_flow_val_hdrs() 236 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && in ice_flow_val_hdrs() 237 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) in ice_flow_val_hdrs() 262 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) in ice_flow_calc_seg_sz() 264 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) in ice_flow_calc_seg_sz() 268 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_TCP) in ice_flow_calc_seg_sz() 270 else if (params->prof->segs[se in ice_flow_calc_seg_sz() 600 ice_flow_find_prof_conds(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, u16 vsi_handle, u32 conds) ice_flow_find_prof_conds() argument 706 ice_flow_add_prof_sync(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, u64 prof_id, struct ice_flow_seg_info *segs, u8 segs_cnt, struct ice_flow_prof **prof) ice_flow_add_prof_sync() argument 881 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, u64 prof_id, struct ice_flow_seg_info *segs, u8 segs_cnt, struct ice_flow_prof **prof) ice_flow_add_prof() argument 1177 ice_flow_set_rss_seg_info(struct ice_flow_seg_info *segs, u64 hash_fields, u32 flow_hdr) ice_flow_set_rss_seg_info() argument 1367 struct ice_flow_seg_info *segs; ice_add_rss_cfg_sync() local [all...] |
/kernel/linux/linux-6.6/net/core/ |
H A D | gso.c | 16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_eth_gso_segment() local 22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment() 28 return segs; in skb_eth_gso_segment() 40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_mac_gso_segment() local 53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment() 61 return segs; in skb_mac_gso_segment() 91 struct sk_buff *segs; in __skb_gso_segment() local 124 segs = skb_mac_gso_segment(skb, features); in __skb_gso_segment() 126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) in __skb_gso_segment() [all...] |
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | udp_offload.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); in __skb_udp_tunnel_segment() local 82 segs = gso_inner_segment(skb, features); in __skb_udp_tunnel_segment() 83 if (IS_ERR_OR_NULL(segs)) { in __skb_udp_tunnel_segment() 89 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment() 93 skb = segs; in __skb_udp_tunnel_segment() 145 return segs; in __skb_udp_tunnel_segment() 155 struct sk_buff *segs = ERR_PTR(-EINVAL); in skb_udp_tunnel_segment() local 177 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, in skb_udp_tunnel_segment() 183 return segs; in skb_udp_tunnel_segment() 214 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) in __udpv4_gso_segment_list_csum() argument 265 struct sk_buff *segs, *seg; __udp_gso_segment() local 367 struct sk_buff *segs = ERR_PTR(-EINVAL); udp4_ufo_fragment() local [all...] |
H A D | tcp_offload.c | 57 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local 89 segs = NULL; in tcp_gso_segment() 98 segs = skb_segment(skb, features); in tcp_gso_segment() 99 if (IS_ERR(segs)) in tcp_gso_segment() 103 segs->ooo_okay = ooo_okay; in tcp_gso_segment() 109 if (skb_is_gso(segs)) in tcp_gso_segment() 110 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 114 skb = segs; in tcp_gso_segment() 119 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 177 return segs; in tcp_gso_segment() [all...] |
H A D | gre_offload.c | 19 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 50 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 51 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 57 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 61 skb = segs; in gre_gso_segment() 113 return segs; in gre_gso_segment()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_flow.c | 624 * @segs: array of one or more packet segments that describe the flow 627 static int ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) in ice_flow_val_hdrs() argument 633 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && in ice_flow_val_hdrs() 634 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) in ice_flow_val_hdrs() 638 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && in ice_flow_val_hdrs() 639 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) in ice_flow_val_hdrs() 667 sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ? in ice_flow_calc_seg_sz() 671 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) in ice_flow_calc_seg_sz() 673 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) in ice_flow_calc_seg_sz() 675 else if (params->prof->segs[se in ice_flow_calc_seg_sz() 1250 ice_flow_find_prof_conds(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, u16 vsi_handle, u32 conds) ice_flow_find_prof_conds() argument 1340 ice_flow_add_prof_sync(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, u64 prof_id, struct ice_flow_seg_info *segs, u8 segs_cnt, struct ice_flow_prof **prof) ice_flow_add_prof_sync() argument 1521 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, u64 prof_id, struct ice_flow_seg_info *segs, u8 segs_cnt, struct ice_flow_prof **prof) ice_flow_add_prof() argument 1869 ice_flow_set_rss_seg_info(struct ice_flow_seg_info *segs, u64 hash_fields, u32 flow_hdr) ice_flow_set_rss_seg_info() argument 2060 struct ice_flow_seg_info *segs; ice_add_rss_cfg_sync() local 2196 struct ice_flow_seg_info *segs; ice_rem_rss_cfg_sync() local [all...] |
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | udp_offload.c | 24 struct sk_buff *segs = ERR_PTR(-EINVAL); in __skb_udp_tunnel_segment() local 87 segs = gso_inner_segment(skb, features); in __skb_udp_tunnel_segment() 88 if (IS_ERR_OR_NULL(segs)) { in __skb_udp_tunnel_segment() 94 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment() 98 skb = segs; in __skb_udp_tunnel_segment() 150 return segs; in __skb_udp_tunnel_segment() 160 struct sk_buff *segs = ERR_PTR(-EINVAL); in skb_udp_tunnel_segment() local 182 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, in skb_udp_tunnel_segment() 188 return segs; in skb_udp_tunnel_segment() 219 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) in __udpv4_gso_segment_list_csum() argument 270 struct sk_buff *segs, *seg; __udp_gso_segment() local 379 struct sk_buff *segs = ERR_PTR(-EINVAL); udp4_ufo_fragment() local [all...] |
H A D | tcp_offload.c | 59 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local 91 segs = NULL; in tcp_gso_segment() 100 segs = skb_segment(skb, features); in tcp_gso_segment() 101 if (IS_ERR(segs)) in tcp_gso_segment() 105 segs->ooo_okay = ooo_okay; in tcp_gso_segment() 111 if (skb_is_gso(segs)) in tcp_gso_segment() 112 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 116 skb = segs; in tcp_gso_segment() 121 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 178 return segs; in tcp_gso_segment() [all...] |
H A D | gre_offload.c | 21 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 58 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 59 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 69 skb = segs; in gre_gso_segment() 119 return segs; in gre_gso_segment()
|
/kernel/linux/linux-5.10/net/mpls/ |
H A D | mpls_gso.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 47 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 48 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 53 skb = segs; in mpls_gso_segment() 69 return segs; in mpls_gso_segment()
|
/kernel/linux/linux-6.6/net/mpls/ |
H A D | mpls_gso.c | 23 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 48 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 49 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 54 skb = segs; in mpls_gso_segment() 70 return segs; in mpls_gso_segment()
|
/kernel/linux/linux-5.10/net/sctp/ |
H A D | offload.c | 37 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 63 segs = NULL; in sctp_gso_segment() 67 segs = skb_segment(skb, features | NETIF_F_HW_CSUM | NETIF_F_SG); in sctp_gso_segment() 68 if (IS_ERR(segs)) in sctp_gso_segment() 73 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 82 return segs; in sctp_gso_segment()
|
/kernel/linux/linux-6.6/net/sctp/ |
H A D | offload.c | 42 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 68 segs = NULL; in sctp_gso_segment() 72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment() 73 if (IS_ERR(segs)) in sctp_gso_segment() 78 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 87 return segs; in sctp_gso_segment()
|
/kernel/linux/linux-5.10/net/nsh/ |
H A D | nsh.c | 79 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 107 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 108 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 114 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 123 return segs; in nsh_gso_segment()
|
/kernel/linux/linux-6.6/net/nsh/ |
H A D | nsh.c | 80 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 108 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 109 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 115 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 124 return segs; in nsh_gso_segment()
|
/kernel/linux/linux-6.6/tools/lib/bpf/ |
H A D | usdt.c | 400 static int parse_elf_segs(Elf *elf, const char *path, struct elf_seg **segs, size_t *seg_cnt) in parse_elf_segs() argument 427 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_elf_segs() 431 *segs = tmp; in parse_elf_segs() 432 seg = *segs + *seg_cnt; in parse_elf_segs() 446 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); in parse_elf_segs() 450 static int parse_vma_segs(int pid, const char *lib_path, struct elf_seg **segs, size_t *seg_cnt) in parse_vma_segs() argument 510 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_vma_segs() 540 find_elf_seg(struct elf_seg *segs, size_t seg_cnt, long virtaddr) find_elf_seg() argument 556 find_vma_seg(struct elf_seg *segs, size_t seg_cnt, long offset) find_vma_seg() argument 583 struct elf_seg *segs = NULL, *vma_segs = NULL; collect_usdt_targets() local [all...] |
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | ip6_offload.c | 76 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 100 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 115 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 116 if (!segs) in ipv6_gso_segment() 120 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 123 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 125 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 140 kfree_skb_list(segs); in ipv6_gso_segment() 155 return segs; in ipv6_gso_segment()
|
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | ip6_offload.c | 78 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 105 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 120 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 121 if (!segs) in ipv6_gso_segment() 125 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 128 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 130 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 145 kfree_skb_list(segs); in ipv6_gso_segment() 160 return segs; in ipv6_gso_segment()
|
/kernel/linux/linux-6.6/include/uapi/linux/ |
H A D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member 45 struct elf64_fdpic_loadseg segs[]; member
|
/kernel/linux/linux-5.10/net/sched/ |
H A D | sch_tbf.c | 194 struct sk_buff *segs, *nskb; in tbf_segment() local 199 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 201 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 205 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 206 skb_mark_not_on_list(segs); in tbf_segment() 207 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment() 208 len += segs->len; in tbf_segment() 209 ret = qdisc_enqueue(segs, in tbf_segment() [all...] |
/kernel/linux/linux-6.6/net/sched/ |
H A D | sch_tbf.c | 209 struct sk_buff *segs, *nskb; in tbf_segment() local 214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 216 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 220 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 221 skb_mark_not_on_list(segs); in tbf_segment() 222 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment() 223 len += segs->len; in tbf_segment() 224 ret = qdisc_enqueue(segs, in tbf_segment() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
H A D | mr.c | 420 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 421 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 610 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 611 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 640 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg() 828 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 829 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 838 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 839 isge->length = mr->map[m]->segs[n].length - off; in rvt_lkey_ok() 935 while (off >= mr->map[m]->segs[ in rvt_rkey_ok() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/ |
H A D | mr.c | 379 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 570 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 599 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg() 787 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 788 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 798 isge->length = mr->map[m]->segs[n].length - off; in rvt_lkey_ok() 894 while (off >= mr->map[m]->segs[ in rvt_rkey_ok() [all...] |
/kernel/linux/linux-5.10/include/rdma/ |
H A D | rdmavt_mr.h | 28 struct rvt_seg segs[RVT_SEGSZ]; member 73 u16 n; /* current index: mr->map[m]->segs[n] */ 133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge() 134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | blk-integrity.h | 69 unsigned int segs) in blk_queue_max_integrity_segments() 71 q->limits.max_integrity_segments = segs; in blk_queue_max_integrity_segments() 154 unsigned int segs) in blk_queue_max_integrity_segments() 68 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) blk_queue_max_integrity_segments() argument 153 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) blk_queue_max_integrity_segments() argument
|