/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt7601u/ |
H A D | dma.c | 29 void *data, u32 seg_len, u32 truesize, struct page *p) in mt7601u_rx_skb_from_seg() 34 skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC); in mt7601u_rx_skb_from_seg() 39 if (!true_len || true_len > seg_len) in mt7601u_rx_skb_from_seg() 77 u32 seg_len, struct page *p) in mt7601u_rx_process_seg() 81 u32 fce_info, truesize = seg_len; in mt7601u_rx_process_seg() 86 fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN); in mt7601u_rx_process_seg() 87 seg_len -= MT_FCE_INFO_LEN; in mt7601u_rx_process_seg() 90 seg_len -= MT_DMA_HDR_LEN; in mt7601u_rx_process_seg() 94 seg_len -= sizeof(struct mt7601u_rxwi); in mt7601u_rx_process_seg() 103 skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesiz in mt7601u_rx_process_seg() 28 mt7601u_rx_skb_from_seg(struct mt7601u_dev *dev, struct mt7601u_rxwi *rxwi, void *data, u32 seg_len, u32 truesize, struct page *p) mt7601u_rx_skb_from_seg() argument 76 mt7601u_rx_process_seg(struct mt7601u_dev *dev, u8 *data, u32 seg_len, struct page *p) mt7601u_rx_process_seg() argument 131 u32 seg_len, data_len = e->urb->actual_length; mt7601u_rx_process_entry() local [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/mediatek/mt7601u/ |
H A D | dma.c | 29 void *data, u32 seg_len, u32 truesize, struct page *p) in mt7601u_rx_skb_from_seg() 34 skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC); in mt7601u_rx_skb_from_seg() 39 if (!true_len || true_len > seg_len) in mt7601u_rx_skb_from_seg() 77 u32 seg_len, struct page *p, in mt7601u_rx_process_seg() 82 u32 fce_info, truesize = seg_len; in mt7601u_rx_process_seg() 87 fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN); in mt7601u_rx_process_seg() 88 seg_len -= MT_FCE_INFO_LEN; in mt7601u_rx_process_seg() 91 seg_len -= MT_DMA_HDR_LEN; in mt7601u_rx_process_seg() 95 seg_len -= sizeof(struct mt7601u_rxwi); in mt7601u_rx_process_seg() 104 skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesiz in mt7601u_rx_process_seg() 28 mt7601u_rx_skb_from_seg(struct mt7601u_dev *dev, struct mt7601u_rxwi *rxwi, void *data, u32 seg_len, u32 truesize, struct page *p) mt7601u_rx_skb_from_seg() argument 76 mt7601u_rx_process_seg(struct mt7601u_dev *dev, u8 *data, u32 seg_len, struct page *p, struct list_head *list) mt7601u_rx_process_seg() argument 136 u32 seg_len, data_len = e->urb->actual_length; mt7601u_rx_process_entry() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_coredump.c | 136 u16 segment_id, u32 *seg_len, in bnxt_hwrm_dbg_coredump_retrieve() 163 *seg_len = info.dest_buf_size; in bnxt_hwrm_dbg_coredump_retrieve() 171 struct coredump_segment_record *seg_rec, u32 seg_len, in bnxt_fill_coredump_seg_hdr() 190 seg_hdr->length = cpu_to_le32(seg_len); in bnxt_fill_coredump_seg_hdr() 315 u32 duration = 0, seg_len = 0; in __bnxt_get_coredump() local 336 &seg_len, buf, buf_len, in __bnxt_get_coredump() 348 bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, seg_record, seg_len, in __bnxt_get_coredump() 354 offset += seg_hdr_len + seg_len; in __bnxt_get_coredump() 357 *dump_len += seg_len; in __bnxt_get_coredump() 135 bnxt_hwrm_dbg_coredump_retrieve(struct bnxt *bp, u16 component_id, u16 segment_id, u32 *seg_len, void *buf, u32 buf_len, u32 offset) bnxt_hwrm_dbg_coredump_retrieve() argument 169 bnxt_fill_coredump_seg_hdr(struct bnxt *bp, struct bnxt_coredump_segment_hdr *seg_hdr, struct coredump_segment_record *seg_rec, u32 seg_len, int status, u32 duration, u32 instance) bnxt_fill_coredump_seg_hdr() argument
|
/kernel/linux/linux-6.6/net/core/ |
H A D | gso.c | 216 * @seg_len: The segmented length (from skb_gso_*_seglen). In the 224 unsigned int seg_len, in skb_gso_size_check() 230 return seg_len <= max_len; in skb_gso_size_check() 233 seg_len -= GSO_BY_FRAGS; in skb_gso_size_check() 236 if (seg_len + skb_headlen(iter) > max_len) in skb_gso_size_check() 223 skb_gso_size_check(const struct sk_buff *skb, unsigned int seg_len, unsigned int max_len) skb_gso_size_check() argument
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_mbox.c | 378 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len() 380 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len() 438 u8 seq_id, seg_len; in recv_mbox_handler() local 443 seg_len = HINIC_MBOX_HEADER_GET(mbox_header, SEG_LEN); in recv_mbox_handler() 446 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler() 449 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler() 683 u16 seg_len) in mbox_copy_send_data() 691 if (seg_len % chk_sz) { in mbox_copy_send_data() 692 memcpy(mbox_max_buf, seg, seg_len); in mbox_copy_send_data() 696 data_len = seg_len; in mbox_copy_send_data() 377 check_mbox_seq_id_and_seg_len(struct hinic_recv_mbox *recv_mbox, u8 seq_id, u8 seg_len) check_mbox_seq_id_and_seg_len() argument 681 mbox_copy_send_data(struct hinic_hwdev *hwdev, struct hinic_send_mbox *mbox, void *seg, u16 seg_len) mbox_copy_send_data() argument 704 write_mbox_msg_attr(struct hinic_mbox_func_to_func *func_to_func, u16 dst_func, u16 dst_aeqn, u16 seg_len, int poll) write_mbox_msg_attr() argument 801 send_mbox_seg(struct hinic_mbox_func_to_func *func_to_func, u64 header, u16 dst_func, void *seg, u16 seg_len, int poll, void *msg_info) send_mbox_seg() argument 852 u16 seg_len = MBOX_SEG_LEN; send_mbox_to_func() local [all...] |
H A D | hinic_hw_mgmt.c | 496 int seq_id, seg_len; in recv_mgmt_msg_handler() local 500 seg_len = HINIC_MSG_HEADER_GET(*header, SEG_LEN); in recv_mgmt_msg_handler() 508 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_mbox.c | 373 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len() 375 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len() 433 u8 seq_id, seg_len; in recv_mbox_handler() local 438 seg_len = HINIC_MBOX_HEADER_GET(mbox_header, SEG_LEN); in recv_mbox_handler() 441 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler() 444 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler() 678 u16 seg_len) in mbox_copy_send_data() 686 if (seg_len % chk_sz) { in mbox_copy_send_data() 687 memcpy(mbox_max_buf, seg, seg_len); in mbox_copy_send_data() 691 data_len = seg_len; in mbox_copy_send_data() 372 check_mbox_seq_id_and_seg_len(struct hinic_recv_mbox *recv_mbox, u8 seq_id, u8 seg_len) check_mbox_seq_id_and_seg_len() argument 676 mbox_copy_send_data(struct hinic_hwdev *hwdev, struct hinic_send_mbox *mbox, void *seg, u16 seg_len) mbox_copy_send_data() argument 699 write_mbox_msg_attr(struct hinic_mbox_func_to_func *func_to_func, u16 dst_func, u16 dst_aeqn, u16 seg_len, int poll) write_mbox_msg_attr() argument 796 send_mbox_seg(struct hinic_mbox_func_to_func *func_to_func, u64 header, u16 dst_func, void *seg, u16 seg_len, int poll, void *msg_info) send_mbox_seg() argument 847 u16 seg_len = MBOX_SEG_LEN; send_mbox_to_func() local [all...] |
H A D | hinic_hw_mgmt.c | 492 int seq_id, seg_len; in recv_mgmt_msg_handler() local 496 seg_len = HINIC_MSG_HEADER_GET(*header, SEG_LEN); in recv_mgmt_msg_handler() 504 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
|
/kernel/linux/linux-5.10/lib/ |
H A D | scatterlist.c | 434 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in __sg_alloc_table_from_pages() local 473 seg_len = 0; in __sg_alloc_table_from_pages() 475 seg_len += PAGE_SIZE; in __sg_alloc_table_from_pages() 476 if (seg_len >= max_segment || in __sg_alloc_table_from_pages() 479 seg_len = 0; in __sg_alloc_table_from_pages() 489 seg_len = 0; in __sg_alloc_table_from_pages() 491 seg_len += PAGE_SIZE; in __sg_alloc_table_from_pages() 492 if (seg_len >= max_segment || in __sg_alloc_table_from_pages()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/google/gve/ |
H A D | gve_desc.h | 30 __be16 seg_len; /* Length of this descriptor's segment */ member 39 __be16 seg_len; member
|
H A D | gve_tx.c | 378 pkt_desc->pkt.seg_len = cpu_to_be16(hlen); in gve_tx_fill_pkt_desc() 393 seg_desc->seg.seg_len = cpu_to_be16(len); in gve_tx_fill_seg_desc()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/google/gve/ |
H A D | gve_desc.h | 32 __be16 seg_len; /* Length of this descriptor's segment */ member 49 __be16 seg_len; member
|
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | lzo.c | 368 u32 seg_len; in lzo_decompress_bio() local 381 seg_len = read_compress_length(kaddr + offset_in_page(cur_in)); in lzo_decompress_bio() 385 if (seg_len > WORKSPACE_CBUF_LENGTH) { in lzo_decompress_bio() 387 * seg_len shouldn't be larger than we have allocated in lzo_decompress_bio() 391 seg_len); in lzo_decompress_bio() 396 copy_compressed_segment(cb, workspace->cbuf, seg_len, &cur_in); in lzo_decompress_bio() 399 ret = lzo1x_decompress_safe(workspace->cbuf, seg_len, in lzo_decompress_bio()
|
/kernel/linux/linux-6.6/drivers/i3c/master/mipi-i3c-hci/ |
H A D | pio.c | 121 unsigned int seg_len, seg_cnt; member 764 p += (ibi->seg_len - ibi->seg_cnt) / 4; in hci_pio_get_ibi_segment() 831 ibi->seg_len = FIELD_GET(IBI_DATA_LENGTH, ibi_status); in hci_pio_prep_new_ibi() 832 ibi->seg_cnt = ibi->seg_len; in hci_pio_prep_new_ibi() 845 if (ibi->seg_len > ibi->max_len) { in hci_pio_prep_new_ibi() 847 ibi->seg_len, ibi->max_len); in hci_pio_prep_new_ibi() 888 ibi->slot->len += ibi->seg_len; in hci_pio_process_ibi() 889 ibi->data_ptr += ibi->seg_len; in hci_pio_process_ibi() 928 ibi->seg_len = FIELD_GET(IBI_DATA_LENGTH, ibi_status); in hci_pio_process_ibi() 929 ibi->seg_cnt = ibi->seg_len; in hci_pio_process_ibi() [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | scatterlist.c | 459 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in sg_alloc_append_table_from_pages() local 501 seg_len = 0; in sg_alloc_append_table_from_pages() 503 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages() 504 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages() 507 seg_len = 0; in sg_alloc_append_table_from_pages() 517 seg_len = 0; in sg_alloc_append_table_from_pages() 519 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages() 520 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 708 int hdr_len, tcp_data, seg_len, pkt_len, offset; in otx2_sq_append_tso() local 730 seg_len = min_t(int, skb_shinfo(skb)->gso_size, tcp_data); in otx2_sq_append_tso() 731 tcp_data -= seg_len; in otx2_sq_append_tso() 741 tso_build_hdr(skb, hdr, &tso, seg_len, tcp_data == 0); in otx2_sq_append_tso() 749 while (seg_len > 0) { in otx2_sq_append_tso() 752 size = min_t(int, tso.size, seg_len); in otx2_sq_append_tso() 762 seg_len -= size; in otx2_sq_append_tso()
|
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_rw.c | 636 unsigned int sge_no, seg_len; in svc_rdma_build_read_segment() local 648 seg_len = min_t(unsigned int, len, in svc_rdma_build_read_segment() 657 seg_len, info->ri_pageoff); in svc_rdma_build_read_segment() 660 info->ri_pageoff += seg_len; in svc_rdma_build_read_segment() 665 len -= seg_len; in svc_rdma_build_read_segment()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_txrx.c | 334 unsigned int seg_len; in ionic_rx_fill() local 364 seg_len = min_t(unsigned int, PAGE_SIZE, len); in ionic_rx_fill() 365 desc->len = cpu_to_le16(seg_len); in ionic_rx_fill() 366 remain_len -= seg_len; in ionic_rx_fill() 381 seg_len = min_t(unsigned int, PAGE_SIZE, remain_len); in ionic_rx_fill() 382 sg_elem->len = cpu_to_le16(seg_len); in ionic_rx_fill() 383 remain_len -= seg_len; in ionic_rx_fill()
|
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_rw.c | 711 unsigned int sge_no, seg_len, len; in svc_rdma_build_read_segment() local 725 seg_len = min_t(unsigned int, len, in svc_rdma_build_read_segment() 732 seg_len, info->ri_pageoff); in svc_rdma_build_read_segment() 735 info->ri_pageoff += seg_len; in svc_rdma_build_read_segment() 740 len -= seg_len; in svc_rdma_build_read_segment()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 884 int hdr_len, tcp_data, seg_len, pkt_len, offset; in otx2_sq_append_tso() local 906 seg_len = min_t(int, skb_shinfo(skb)->gso_size, tcp_data); in otx2_sq_append_tso() 907 tcp_data -= seg_len; in otx2_sq_append_tso() 917 tso_build_hdr(skb, hdr, &tso, seg_len, tcp_data == 0); in otx2_sq_append_tso() 925 while (seg_len > 0) { in otx2_sq_append_tso() 928 size = min_t(int, tso.size, seg_len); in otx2_sq_append_tso() 938 seg_len -= size; in otx2_sq_append_tso()
|
/kernel/linux/linux-5.10/drivers/bus/mhi/host/ |
H A D | boot.c | 311 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table() 425 !mhi_cntrl->seg_len))) { in mhi_fw_load_handler() 427 "No firmware image defined or !sbl_size || !seg_len\n"); in mhi_fw_load_handler()
|
/kernel/linux/linux-6.6/drivers/bus/mhi/host/ |
H A D | boot.c | 317 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table() 442 !mhi_cntrl->seg_len))) { in mhi_fw_load_handler() 444 "No firmware image defined or !sbl_size || !seg_len\n"); in mhi_fw_load_handler()
|
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | ip6_output.c | 269 int seg_len = skb->len; in ip6_xmit() local 286 seg_len += opt->opt_nflen + opt->opt_flen; in ip6_xmit() 296 if (unlikely(seg_len > IPV6_MAXPLEN)) { in ip6_xmit() 303 hop_jumbo->jumbo_payload_len = htonl(seg_len + hoplen); in ip6_xmit() 306 seg_len = 0; in ip6_xmit() 325 hdr->payload_len = htons(seg_len); in ip6_xmit()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | tid_rdma.c | 1086 u32 length = flow->req->seg_len; in kern_find_pages() 1110 flow->length = flow->req->seg_len - length; in kern_find_pages() 1166 * segment. All segments are of length flow->req->seg_len. 1434 * function uses qp, rcd and seg_len members of @req. In the absence of errors, 1742 req->cur_seg * req->seg_len + flow->sent); in hfi1_build_tid_rdma_read_packet() 1968 req->seg_len = qpriv->tid_rdma.local.max_len; in tid_rdma_rcv_read_request() 3098 rvt_skip_sge(&qpriv->tid_ss, (req->cur_seg * req->seg_len) + in hfi1_tid_rdma_restart_req() 3318 priv->tid_req.seg_len = in setup_tid_rdma_wqe() 3321 DIV_ROUND_UP(wqe->length, priv->tid_req.seg_len); in setup_tid_rdma_wqe() 3530 npkts = rvt_div_round_up_mtu(qp, req->seg_len); in hfi1_tid_write_alloc_resources() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | tid_rdma.c | 1093 u32 length = flow->req->seg_len; in kern_find_pages() 1117 flow->length = flow->req->seg_len - length; in kern_find_pages() 1173 * segment. All segments are of length flow->req->seg_len. 1441 * function uses qp, rcd and seg_len members of @req. In the absence of errors, 1749 req->cur_seg * req->seg_len + flow->sent); in hfi1_build_tid_rdma_read_packet() 1975 req->seg_len = qpriv->tid_rdma.local.max_len; in tid_rdma_rcv_read_request() 3107 rvt_skip_sge(&qpriv->tid_ss, (req->cur_seg * req->seg_len) + in hfi1_tid_rdma_restart_req() 3330 priv->tid_req.seg_len = in setup_tid_rdma_wqe() 3333 DIV_ROUND_UP(wqe->length, priv->tid_req.seg_len); in setup_tid_rdma_wqe() 3542 npkts = rvt_div_round_up_mtu(qp, req->seg_len); in hfi1_tid_write_alloc_resources() [all...] |