Home
last modified time | relevance | path

Searched refs:seg_len (Results 1 - 25 of 51) sorted by relevance

123

/kernel/linux/linux-5.10/drivers/net/wireless/mediatek/mt7601u/
H A Ddma.c29 void *data, u32 seg_len, u32 truesize, struct page *p) in mt7601u_rx_skb_from_seg()
34 skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC); in mt7601u_rx_skb_from_seg()
39 if (!true_len || true_len > seg_len) in mt7601u_rx_skb_from_seg()
77 u32 seg_len, struct page *p) in mt7601u_rx_process_seg()
81 u32 fce_info, truesize = seg_len; in mt7601u_rx_process_seg()
86 fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN); in mt7601u_rx_process_seg()
87 seg_len -= MT_FCE_INFO_LEN; in mt7601u_rx_process_seg()
90 seg_len -= MT_DMA_HDR_LEN; in mt7601u_rx_process_seg()
94 seg_len -= sizeof(struct mt7601u_rxwi); in mt7601u_rx_process_seg()
103 skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesiz in mt7601u_rx_process_seg()
28 mt7601u_rx_skb_from_seg(struct mt7601u_dev *dev, struct mt7601u_rxwi *rxwi, void *data, u32 seg_len, u32 truesize, struct page *p) mt7601u_rx_skb_from_seg() argument
76 mt7601u_rx_process_seg(struct mt7601u_dev *dev, u8 *data, u32 seg_len, struct page *p) mt7601u_rx_process_seg() argument
131 u32 seg_len, data_len = e->urb->actual_length; mt7601u_rx_process_entry() local
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/mediatek/mt7601u/
H A Ddma.c29 void *data, u32 seg_len, u32 truesize, struct page *p) in mt7601u_rx_skb_from_seg()
34 skb = alloc_skb(p ? 128 : seg_len, GFP_ATOMIC); in mt7601u_rx_skb_from_seg()
39 if (!true_len || true_len > seg_len) in mt7601u_rx_skb_from_seg()
77 u32 seg_len, struct page *p, in mt7601u_rx_process_seg()
82 u32 fce_info, truesize = seg_len; in mt7601u_rx_process_seg()
87 fce_info = get_unaligned_le32(data + seg_len - MT_FCE_INFO_LEN); in mt7601u_rx_process_seg()
88 seg_len -= MT_FCE_INFO_LEN; in mt7601u_rx_process_seg()
91 seg_len -= MT_DMA_HDR_LEN; in mt7601u_rx_process_seg()
95 seg_len -= sizeof(struct mt7601u_rxwi); in mt7601u_rx_process_seg()
104 skb = mt7601u_rx_skb_from_seg(dev, rxwi, data, seg_len, truesiz in mt7601u_rx_process_seg()
28 mt7601u_rx_skb_from_seg(struct mt7601u_dev *dev, struct mt7601u_rxwi *rxwi, void *data, u32 seg_len, u32 truesize, struct page *p) mt7601u_rx_skb_from_seg() argument
76 mt7601u_rx_process_seg(struct mt7601u_dev *dev, u8 *data, u32 seg_len, struct page *p, struct list_head *list) mt7601u_rx_process_seg() argument
136 u32 seg_len, data_len = e->urb->actual_length; mt7601u_rx_process_entry() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_coredump.c136 u16 segment_id, u32 *seg_len, in bnxt_hwrm_dbg_coredump_retrieve()
163 *seg_len = info.dest_buf_size; in bnxt_hwrm_dbg_coredump_retrieve()
171 struct coredump_segment_record *seg_rec, u32 seg_len, in bnxt_fill_coredump_seg_hdr()
190 seg_hdr->length = cpu_to_le32(seg_len); in bnxt_fill_coredump_seg_hdr()
315 u32 duration = 0, seg_len = 0; in __bnxt_get_coredump() local
336 &seg_len, buf, buf_len, in __bnxt_get_coredump()
348 bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, seg_record, seg_len, in __bnxt_get_coredump()
354 offset += seg_hdr_len + seg_len; in __bnxt_get_coredump()
357 *dump_len += seg_len; in __bnxt_get_coredump()
135 bnxt_hwrm_dbg_coredump_retrieve(struct bnxt *bp, u16 component_id, u16 segment_id, u32 *seg_len, void *buf, u32 buf_len, u32 offset) bnxt_hwrm_dbg_coredump_retrieve() argument
169 bnxt_fill_coredump_seg_hdr(struct bnxt *bp, struct bnxt_coredump_segment_hdr *seg_hdr, struct coredump_segment_record *seg_rec, u32 seg_len, int status, u32 duration, u32 instance) bnxt_fill_coredump_seg_hdr() argument
/kernel/linux/linux-6.6/net/core/
H A Dgso.c216 * @seg_len: The segmented length (from skb_gso_*_seglen). In the
224 unsigned int seg_len, in skb_gso_size_check()
230 return seg_len <= max_len; in skb_gso_size_check()
233 seg_len -= GSO_BY_FRAGS; in skb_gso_size_check()
236 if (seg_len + skb_headlen(iter) > max_len) in skb_gso_size_check()
223 skb_gso_size_check(const struct sk_buff *skb, unsigned int seg_len, unsigned int max_len) skb_gso_size_check() argument
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_mbox.c378 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len()
380 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len()
438 u8 seq_id, seg_len; in recv_mbox_handler() local
443 seg_len = HINIC_MBOX_HEADER_GET(mbox_header, SEG_LEN); in recv_mbox_handler()
446 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler()
449 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler()
683 u16 seg_len) in mbox_copy_send_data()
691 if (seg_len % chk_sz) { in mbox_copy_send_data()
692 memcpy(mbox_max_buf, seg, seg_len); in mbox_copy_send_data()
696 data_len = seg_len; in mbox_copy_send_data()
377 check_mbox_seq_id_and_seg_len(struct hinic_recv_mbox *recv_mbox, u8 seq_id, u8 seg_len) check_mbox_seq_id_and_seg_len() argument
681 mbox_copy_send_data(struct hinic_hwdev *hwdev, struct hinic_send_mbox *mbox, void *seg, u16 seg_len) mbox_copy_send_data() argument
704 write_mbox_msg_attr(struct hinic_mbox_func_to_func *func_to_func, u16 dst_func, u16 dst_aeqn, u16 seg_len, int poll) write_mbox_msg_attr() argument
801 send_mbox_seg(struct hinic_mbox_func_to_func *func_to_func, u64 header, u16 dst_func, void *seg, u16 seg_len, int poll, void *msg_info) send_mbox_seg() argument
852 u16 seg_len = MBOX_SEG_LEN; send_mbox_to_func() local
[all...]
H A Dhinic_hw_mgmt.c496 int seq_id, seg_len; in recv_mgmt_msg_handler() local
500 seg_len = HINIC_MSG_HEADER_GET(*header, SEG_LEN); in recv_mgmt_msg_handler()
508 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_mbox.c373 u8 seq_id, u8 seg_len) in check_mbox_seq_id_and_seg_len()
375 if (seq_id > SEQ_ID_MAX_VAL || seg_len > MBOX_SEG_LEN) in check_mbox_seq_id_and_seg_len()
433 u8 seq_id, seg_len; in recv_mbox_handler() local
438 seg_len = HINIC_MBOX_HEADER_GET(mbox_header, SEG_LEN); in recv_mbox_handler()
441 if (!check_mbox_seq_id_and_seg_len(recv_mbox, seq_id, seg_len)) { in recv_mbox_handler()
444 src_func_idx, recv_mbox->seq_id, seq_id, seg_len); in recv_mbox_handler()
678 u16 seg_len) in mbox_copy_send_data()
686 if (seg_len % chk_sz) { in mbox_copy_send_data()
687 memcpy(mbox_max_buf, seg, seg_len); in mbox_copy_send_data()
691 data_len = seg_len; in mbox_copy_send_data()
372 check_mbox_seq_id_and_seg_len(struct hinic_recv_mbox *recv_mbox, u8 seq_id, u8 seg_len) check_mbox_seq_id_and_seg_len() argument
676 mbox_copy_send_data(struct hinic_hwdev *hwdev, struct hinic_send_mbox *mbox, void *seg, u16 seg_len) mbox_copy_send_data() argument
699 write_mbox_msg_attr(struct hinic_mbox_func_to_func *func_to_func, u16 dst_func, u16 dst_aeqn, u16 seg_len, int poll) write_mbox_msg_attr() argument
796 send_mbox_seg(struct hinic_mbox_func_to_func *func_to_func, u64 header, u16 dst_func, void *seg, u16 seg_len, int poll, void *msg_info) send_mbox_seg() argument
847 u16 seg_len = MBOX_SEG_LEN; send_mbox_to_func() local
[all...]
H A Dhinic_hw_mgmt.c492 int seq_id, seg_len; in recv_mgmt_msg_handler() local
496 seg_len = HINIC_MSG_HEADER_GET(*header, SEG_LEN); in recv_mgmt_msg_handler()
504 memcpy(recv_msg->msg + seq_id * SEGMENT_LEN, msg_body, seg_len); in recv_mgmt_msg_handler()
/kernel/linux/linux-5.10/lib/
H A Dscatterlist.c434 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in __sg_alloc_table_from_pages() local
473 seg_len = 0; in __sg_alloc_table_from_pages()
475 seg_len += PAGE_SIZE; in __sg_alloc_table_from_pages()
476 if (seg_len >= max_segment || in __sg_alloc_table_from_pages()
479 seg_len = 0; in __sg_alloc_table_from_pages()
489 seg_len = 0; in __sg_alloc_table_from_pages()
491 seg_len += PAGE_SIZE; in __sg_alloc_table_from_pages()
492 if (seg_len >= max_segment || in __sg_alloc_table_from_pages()
/kernel/linux/linux-5.10/drivers/net/ethernet/google/gve/
H A Dgve_desc.h30 __be16 seg_len; /* Length of this descriptor's segment */ member
39 __be16 seg_len; member
H A Dgve_tx.c378 pkt_desc->pkt.seg_len = cpu_to_be16(hlen); in gve_tx_fill_pkt_desc()
393 seg_desc->seg.seg_len = cpu_to_be16(len); in gve_tx_fill_seg_desc()
/kernel/linux/linux-6.6/drivers/net/ethernet/google/gve/
H A Dgve_desc.h32 __be16 seg_len; /* Length of this descriptor's segment */ member
49 __be16 seg_len; member
/kernel/linux/linux-6.6/fs/btrfs/
H A Dlzo.c368 u32 seg_len; in lzo_decompress_bio() local
381 seg_len = read_compress_length(kaddr + offset_in_page(cur_in)); in lzo_decompress_bio()
385 if (seg_len > WORKSPACE_CBUF_LENGTH) { in lzo_decompress_bio()
387 * seg_len shouldn't be larger than we have allocated in lzo_decompress_bio()
391 seg_len); in lzo_decompress_bio()
396 copy_compressed_segment(cb, workspace->cbuf, seg_len, &cur_in); in lzo_decompress_bio()
399 ret = lzo1x_decompress_safe(workspace->cbuf, seg_len, in lzo_decompress_bio()
/kernel/linux/linux-6.6/drivers/i3c/master/mipi-i3c-hci/
H A Dpio.c121 unsigned int seg_len, seg_cnt; member
764 p += (ibi->seg_len - ibi->seg_cnt) / 4; in hci_pio_get_ibi_segment()
831 ibi->seg_len = FIELD_GET(IBI_DATA_LENGTH, ibi_status); in hci_pio_prep_new_ibi()
832 ibi->seg_cnt = ibi->seg_len; in hci_pio_prep_new_ibi()
845 if (ibi->seg_len > ibi->max_len) { in hci_pio_prep_new_ibi()
847 ibi->seg_len, ibi->max_len); in hci_pio_prep_new_ibi()
888 ibi->slot->len += ibi->seg_len; in hci_pio_process_ibi()
889 ibi->data_ptr += ibi->seg_len; in hci_pio_process_ibi()
928 ibi->seg_len = FIELD_GET(IBI_DATA_LENGTH, ibi_status); in hci_pio_process_ibi()
929 ibi->seg_cnt = ibi->seg_len; in hci_pio_process_ibi()
[all...]
/kernel/linux/linux-6.6/lib/
H A Dscatterlist.c459 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in sg_alloc_append_table_from_pages() local
501 seg_len = 0; in sg_alloc_append_table_from_pages()
503 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages()
504 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages()
507 seg_len = 0; in sg_alloc_append_table_from_pages()
517 seg_len = 0; in sg_alloc_append_table_from_pages()
519 seg_len += PAGE_SIZE; in sg_alloc_append_table_from_pages()
520 if (seg_len >= max_segment || in sg_alloc_append_table_from_pages()
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c708 int hdr_len, tcp_data, seg_len, pkt_len, offset; in otx2_sq_append_tso() local
730 seg_len = min_t(int, skb_shinfo(skb)->gso_size, tcp_data); in otx2_sq_append_tso()
731 tcp_data -= seg_len; in otx2_sq_append_tso()
741 tso_build_hdr(skb, hdr, &tso, seg_len, tcp_data == 0); in otx2_sq_append_tso()
749 while (seg_len > 0) { in otx2_sq_append_tso()
752 size = min_t(int, tso.size, seg_len); in otx2_sq_append_tso()
762 seg_len -= size; in otx2_sq_append_tso()
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/
H A Dsvc_rdma_rw.c636 unsigned int sge_no, seg_len; in svc_rdma_build_read_segment() local
648 seg_len = min_t(unsigned int, len, in svc_rdma_build_read_segment()
657 seg_len, info->ri_pageoff); in svc_rdma_build_read_segment()
660 info->ri_pageoff += seg_len; in svc_rdma_build_read_segment()
665 len -= seg_len; in svc_rdma_build_read_segment()
/kernel/linux/linux-5.10/drivers/net/ethernet/pensando/ionic/
H A Dionic_txrx.c334 unsigned int seg_len; in ionic_rx_fill() local
364 seg_len = min_t(unsigned int, PAGE_SIZE, len); in ionic_rx_fill()
365 desc->len = cpu_to_le16(seg_len); in ionic_rx_fill()
366 remain_len -= seg_len; in ionic_rx_fill()
381 seg_len = min_t(unsigned int, PAGE_SIZE, remain_len); in ionic_rx_fill()
382 sg_elem->len = cpu_to_le16(seg_len); in ionic_rx_fill()
383 remain_len -= seg_len; in ionic_rx_fill()
/kernel/linux/linux-6.6/net/sunrpc/xprtrdma/
H A Dsvc_rdma_rw.c711 unsigned int sge_no, seg_len, len; in svc_rdma_build_read_segment() local
725 seg_len = min_t(unsigned int, len, in svc_rdma_build_read_segment()
732 seg_len, info->ri_pageoff); in svc_rdma_build_read_segment()
735 info->ri_pageoff += seg_len; in svc_rdma_build_read_segment()
740 len -= seg_len; in svc_rdma_build_read_segment()
/kernel/linux/linux-6.6/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c884 int hdr_len, tcp_data, seg_len, pkt_len, offset; in otx2_sq_append_tso() local
906 seg_len = min_t(int, skb_shinfo(skb)->gso_size, tcp_data); in otx2_sq_append_tso()
907 tcp_data -= seg_len; in otx2_sq_append_tso()
917 tso_build_hdr(skb, hdr, &tso, seg_len, tcp_data == 0); in otx2_sq_append_tso()
925 while (seg_len > 0) { in otx2_sq_append_tso()
928 size = min_t(int, tso.size, seg_len); in otx2_sq_append_tso()
938 seg_len -= size; in otx2_sq_append_tso()
/kernel/linux/linux-5.10/drivers/bus/mhi/host/
H A Dboot.c311 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table()
425 !mhi_cntrl->seg_len))) { in mhi_fw_load_handler()
427 "No firmware image defined or !sbl_size || !seg_len\n"); in mhi_fw_load_handler()
/kernel/linux/linux-6.6/drivers/bus/mhi/host/
H A Dboot.c317 size_t seg_size = mhi_cntrl->seg_len; in mhi_alloc_bhie_table()
442 !mhi_cntrl->seg_len))) { in mhi_fw_load_handler()
444 "No firmware image defined or !sbl_size || !seg_len\n"); in mhi_fw_load_handler()
/kernel/linux/linux-6.6/net/ipv6/
H A Dip6_output.c269 int seg_len = skb->len; in ip6_xmit() local
286 seg_len += opt->opt_nflen + opt->opt_flen; in ip6_xmit()
296 if (unlikely(seg_len > IPV6_MAXPLEN)) { in ip6_xmit()
303 hop_jumbo->jumbo_payload_len = htonl(seg_len + hoplen); in ip6_xmit()
306 seg_len = 0; in ip6_xmit()
325 hdr->payload_len = htons(seg_len); in ip6_xmit()
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
H A Dtid_rdma.c1086 u32 length = flow->req->seg_len; in kern_find_pages()
1110 flow->length = flow->req->seg_len - length; in kern_find_pages()
1166 * segment. All segments are of length flow->req->seg_len.
1434 * function uses qp, rcd and seg_len members of @req. In the absence of errors,
1742 req->cur_seg * req->seg_len + flow->sent); in hfi1_build_tid_rdma_read_packet()
1968 req->seg_len = qpriv->tid_rdma.local.max_len; in tid_rdma_rcv_read_request()
3098 rvt_skip_sge(&qpriv->tid_ss, (req->cur_seg * req->seg_len) + in hfi1_tid_rdma_restart_req()
3318 priv->tid_req.seg_len = in setup_tid_rdma_wqe()
3321 DIV_ROUND_UP(wqe->length, priv->tid_req.seg_len); in setup_tid_rdma_wqe()
3530 npkts = rvt_div_round_up_mtu(qp, req->seg_len); in hfi1_tid_write_alloc_resources()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/
H A Dtid_rdma.c1093 u32 length = flow->req->seg_len; in kern_find_pages()
1117 flow->length = flow->req->seg_len - length; in kern_find_pages()
1173 * segment. All segments are of length flow->req->seg_len.
1441 * function uses qp, rcd and seg_len members of @req. In the absence of errors,
1749 req->cur_seg * req->seg_len + flow->sent); in hfi1_build_tid_rdma_read_packet()
1975 req->seg_len = qpriv->tid_rdma.local.max_len; in tid_rdma_rcv_read_request()
3107 rvt_skip_sge(&qpriv->tid_ss, (req->cur_seg * req->seg_len) + in hfi1_tid_rdma_restart_req()
3330 priv->tid_req.seg_len = in setup_tid_rdma_wqe()
3333 DIV_ROUND_UP(wqe->length, priv->tid_req.seg_len); in setup_tid_rdma_wqe()
3542 npkts = rvt_div_round_up_mtu(qp, req->seg_len); in hfi1_tid_write_alloc_resources()
[all...]

Completed in 36 milliseconds

123