/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 490 const struct ib_send_wr *wr, u8 *len16) in build_rdma_send() 551 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_send() 557 const struct ib_send_wr *wr, u8 *len16) in build_rdma_write() 602 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_write() 618 const struct ib_send_wr *wr, u8 *len16) in build_rdma_write_cmpl() 657 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_write_cmpl() 661 u8 *len16) in build_rdma_read() 686 *len16 = DIV_ROUND_UP(sizeof(wqe->read), 16); in build_rdma_read() 699 u8 len16; in post_write_cmpl() local 708 build_rdma_write_cmpl(&qhp->wq.sq, &wqe->write_cmpl, wr, &len16); in post_write_cmpl() 489 build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_send() argument 556 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_write() argument 616 build_rdma_write_cmpl(struct t4_sq *sq, struct fw_ri_rdma_write_cmpl_wr *wcwr, const struct ib_send_wr *wr, u8 *len16) build_rdma_write_cmpl() argument 660 build_rdma_read(union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_read() argument 759 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, const struct ib_recv_wr *wr, u8 *len16) build_rdma_recv() argument 774 build_srq_recv(union t4_recv_wr *wqe, const struct ib_recv_wr *wr, u8 *len16) build_srq_recv() argument 788 build_tpte_memreg(struct fw_ri_fr_nsmr_tpte_wr *fr, const struct ib_reg_wr *wr, struct c4iw_mr *mhp, u8 *len16) build_tpte_memreg() argument 820 build_memreg(struct t4_sq *sq, union t4_wr *wqe, const struct ib_reg_wr *wr, struct c4iw_mr *mhp, u8 *len16, bool dsgl_supported) build_memreg() argument 884 build_inv_stag(union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_inv_stag() argument 1081 u8 len16 = 0; c4iw_post_send() local 1268 u8 len16 = 0; c4iw_post_receive() local 1341 defer_srq_wr(struct t4_srq *srq, union t4_recv_wr *wqe, u64 wr_id, u8 len16) defer_srq_wr() argument 1363 u8 len16 = 0; c4iw_post_srq_recv() local 2654 c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16) c4iw_copy_wr_to_srq() argument [all...] |
H A D | t4.h | 120 enum fw_wr_opcodes opcode, u8 flags, u8 len16) in init_wr_hdr() 128 wqe->send.len16 = len16; in init_wr_hdr() 395 u8 len16; member 430 static inline void t4_srq_produce(struct t4_srq *srq, u8 len16) in t4_srq_produce() argument 435 srq->wq_pidx += DIV_ROUND_UP(len16 * 16, T4_EQ_ENTRY_SIZE); in t4_srq_produce() 500 static inline void t4_rq_produce(struct t4_wq *wq, u8 len16) in t4_rq_produce() argument 505 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce() 547 static inline void t4_sq_produce(struct t4_wq *wq, u8 len16) in t4_sq_produce() argument 552 wq->sq.wq_pidx += DIV_ROUND_UP(len16*1 in t4_sq_produce() 119 init_wr_hdr(union t4_wr *wqe, u16 wrid, enum fw_wr_opcodes opcode, u8 flags, u8 len16) init_wr_hdr() argument 592 t4_ring_srq_db(struct t4_srq *srq, u16 inc, u8 len16, union t4_recv_wr *wqe) t4_ring_srq_db() argument [all...] |
H A D | t4fw_ri_api.h | 550 __u8 len16; member 578 __u8 len16; member 603 __u8 len16; member 634 __u8 len16; member 651 __u8 len16; member 660 __u8 len16; member 698 __u8 len16; member 735 __u8 len16; member 747 __u8 len16; member
|
H A D | cq.c | 478 c4iw_copy_wr_to_srq(srq, &pwr->wqe, pwr->len16); in post_pending_srq_wrs() 480 t4_srq_produce(srq, pwr->len16); in post_pending_srq_wrs() 481 idx += DIV_ROUND_UP(pwr->len16 * 16, T4_EQ_ENTRY_SIZE); in post_pending_srq_wrs() 485 t4_ring_srq_db(srq, idx, pwr->len16, &pwr->wqe); in post_pending_srq_wrs()
|
H A D | mem.c | 94 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), 16)); in _c4iw_write_mem_dma_aligned() 159 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), in _c4iw_write_mem_inline()
|
H A D | iw_cxgb4.h | 1049 void c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16);
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 490 const struct ib_send_wr *wr, u8 *len16) in build_rdma_send() 551 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_send() 557 const struct ib_send_wr *wr, u8 *len16) in build_rdma_write() 602 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_write() 618 const struct ib_send_wr *wr, u8 *len16) in build_rdma_write_cmpl() 657 *len16 = DIV_ROUND_UP(size, 16); in build_rdma_write_cmpl() 661 u8 *len16) in build_rdma_read() 686 *len16 = DIV_ROUND_UP(sizeof(wqe->read), 16); in build_rdma_read() 699 u8 len16; in post_write_cmpl() local 708 build_rdma_write_cmpl(&qhp->wq.sq, &wqe->write_cmpl, wr, &len16); in post_write_cmpl() 489 build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_send() argument 556 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_write() argument 616 build_rdma_write_cmpl(struct t4_sq *sq, struct fw_ri_rdma_write_cmpl_wr *wcwr, const struct ib_send_wr *wr, u8 *len16) build_rdma_write_cmpl() argument 660 build_rdma_read(union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_rdma_read() argument 759 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, const struct ib_recv_wr *wr, u8 *len16) build_rdma_recv() argument 774 build_srq_recv(union t4_recv_wr *wqe, const struct ib_recv_wr *wr, u8 *len16) build_srq_recv() argument 788 build_tpte_memreg(struct fw_ri_fr_nsmr_tpte_wr *fr, const struct ib_reg_wr *wr, struct c4iw_mr *mhp, u8 *len16) build_tpte_memreg() argument 820 build_memreg(struct t4_sq *sq, union t4_wr *wqe, const struct ib_reg_wr *wr, struct c4iw_mr *mhp, u8 *len16, bool dsgl_supported) build_memreg() argument 884 build_inv_stag(union t4_wr *wqe, const struct ib_send_wr *wr, u8 *len16) build_inv_stag() argument 1081 u8 len16 = 0; c4iw_post_send() local 1268 u8 len16 = 0; c4iw_post_receive() local 1341 defer_srq_wr(struct t4_srq *srq, union t4_recv_wr *wqe, u64 wr_id, u8 len16) defer_srq_wr() argument 1363 u8 len16 = 0; c4iw_post_srq_recv() local 2645 c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16) c4iw_copy_wr_to_srq() argument [all...] |
H A D | t4.h | 120 enum fw_wr_opcodes opcode, u8 flags, u8 len16) in init_wr_hdr() 128 wqe->send.len16 = len16; in init_wr_hdr() 395 u8 len16; member 430 static inline void t4_srq_produce(struct t4_srq *srq, u8 len16) in t4_srq_produce() argument 435 srq->wq_pidx += DIV_ROUND_UP(len16 * 16, T4_EQ_ENTRY_SIZE); in t4_srq_produce() 495 static inline void t4_rq_produce(struct t4_wq *wq, u8 len16) in t4_rq_produce() argument 500 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce() 537 static inline void t4_sq_produce(struct t4_wq *wq, u8 len16) in t4_sq_produce() argument 542 wq->sq.wq_pidx += DIV_ROUND_UP(len16*1 in t4_sq_produce() 119 init_wr_hdr(union t4_wr *wqe, u16 wrid, enum fw_wr_opcodes opcode, u8 flags, u8 len16) init_wr_hdr() argument 582 t4_ring_srq_db(struct t4_srq *srq, u16 inc, u8 len16, union t4_recv_wr *wqe) t4_ring_srq_db() argument [all...] |
H A D | t4fw_ri_api.h | 542 __u8 len16; member 568 __u8 len16; member 591 __u8 len16; member 620 __u8 len16; member 637 __u8 len16; member 646 __u8 len16; member 684 __u8 len16; member 721 __u8 len16; member 733 __u8 len16; member
|
H A D | cq.c | 478 c4iw_copy_wr_to_srq(srq, &pwr->wqe, pwr->len16); in post_pending_srq_wrs() 480 t4_srq_produce(srq, pwr->len16); in post_pending_srq_wrs() 481 idx += DIV_ROUND_UP(pwr->len16 * 16, T4_EQ_ENTRY_SIZE); in post_pending_srq_wrs() 485 t4_ring_srq_db(srq, idx, pwr->len16, &pwr->wqe); in post_pending_srq_wrs()
|
H A D | mem.c | 94 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), 16)); in _c4iw_write_mem_dma_aligned() 159 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), in _c4iw_write_mem_inline()
|
H A D | iw_cxgb4.h | 1035 void c4iw_copy_wr_to_srq(struct t4_srq *srq, union t4_recv_wr *wqe, u8 len16);
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
H A D | chcr_ktls.c | 1007 int len16, pktlen; in chcr_ktls_write_tcp_options() local 1024 len16 = DIV_ROUND_UP(sizeof(*wr) + ctrl, 16); in chcr_ktls_write_tcp_options() 1026 ndesc = DIV_ROUND_UP(len16, 4); in chcr_ktls_write_tcp_options() 1046 wr->equiq_to_len16 = htonl(wr_mid | FW_WR_LEN16_V(len16)); in chcr_ktls_write_tcp_options() 1113 u32 len16, wr_mid = 0, flits = 0, ndesc, cipher_start; in chcr_ktls_xmit_wr_complete() local 1162 /* WR will need len16 */ in chcr_ktls_xmit_wr_complete() 1163 len16 = DIV_ROUND_UP(flits, 2); in chcr_ktls_xmit_wr_complete() 1165 wr->flowid_len16 = htonl(wr_mid | FW_WR_LEN16_V(len16)); in chcr_ktls_xmit_wr_complete() 1174 ulptx->len = htonl(len16 - 1); in chcr_ktls_xmit_wr_complete() 1288 u32 len16, wr_mi in chcr_ktls_xmit_wr_short() local 1474 int credits, left, len16, last_desc; chcr_ktls_tx_plaintxt() local 1580 u32 ctrl, iplen, maclen, wr_mid = 0, len16; chcr_ktls_tunnel_pkt() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
H A D | chcr_ktls.c | 1001 int len16, pktlen; in chcr_ktls_write_tcp_options() local 1018 len16 = DIV_ROUND_UP(sizeof(*wr) + ctrl, 16); in chcr_ktls_write_tcp_options() 1020 ndesc = DIV_ROUND_UP(len16, 4); in chcr_ktls_write_tcp_options() 1040 wr->equiq_to_len16 = htonl(wr_mid | FW_WR_LEN16_V(len16)); in chcr_ktls_write_tcp_options() 1107 u32 len16, wr_mid = 0, flits = 0, ndesc, cipher_start; in chcr_ktls_xmit_wr_complete() local 1156 /* WR will need len16 */ in chcr_ktls_xmit_wr_complete() 1157 len16 = DIV_ROUND_UP(flits, 2); in chcr_ktls_xmit_wr_complete() 1159 wr->flowid_len16 = htonl(wr_mid | FW_WR_LEN16_V(len16)); in chcr_ktls_xmit_wr_complete() 1168 ulptx->len = htonl(len16 - 1); in chcr_ktls_xmit_wr_complete() 1282 u32 len16, wr_mi in chcr_ktls_xmit_wr_short() local 1468 int credits, left, len16, last_desc; chcr_ktls_tx_plaintxt() local 1573 u32 ctrl, iplen, maclen, wr_mid = 0, len16; chcr_ktls_tunnel_pkt() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | t4vf_hw.c | 627 size_t len16; in t4vf_query_params() local 636 len16 = DIV_ROUND_UP(offsetof(struct fw_params_cmd, in t4vf_query_params() 638 cmd.retval_len16 = cpu_to_be32(FW_CMD_LEN16_V(len16)); in t4vf_query_params() 665 size_t len16; in t4vf_set_params() local 674 len16 = DIV_ROUND_UP(offsetof(struct fw_params_cmd, in t4vf_set_params() 676 cmd.retval_len16 = cpu_to_be32(FW_CMD_LEN16_V(len16)); in t4vf_set_params() 1543 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, in t4vf_alloc_mac_filt() local 1556 FW_CMD_LEN16_V(len16)); in t4vf_alloc_mac_filt() 1627 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, in t4vf_free_mac_filt() local 1640 FW_CMD_LEN16_V(len16)); in t4vf_free_mac_filt() 1695 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, t4vf_change_mac() local 1740 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, t4vf_set_addr_hash() local 1782 size_t len16 = DIV_ROUND_UP(len, 16); t4vf_get_port_stats() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | t4vf_hw.c | 628 size_t len16; in t4vf_query_params() local 637 len16 = DIV_ROUND_UP(offsetof(struct fw_params_cmd, in t4vf_query_params() 639 cmd.retval_len16 = cpu_to_be32(FW_CMD_LEN16_V(len16)); in t4vf_query_params() 666 size_t len16; in t4vf_set_params() local 675 len16 = DIV_ROUND_UP(offsetof(struct fw_params_cmd, in t4vf_set_params() 677 cmd.retval_len16 = cpu_to_be32(FW_CMD_LEN16_V(len16)); in t4vf_set_params() 1544 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, in t4vf_alloc_mac_filt() local 1557 FW_CMD_LEN16_V(len16)); in t4vf_alloc_mac_filt() 1628 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, in t4vf_free_mac_filt() local 1641 FW_CMD_LEN16_V(len16)); in t4vf_free_mac_filt() 1696 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, t4vf_change_mac() local 1741 size_t len16 = DIV_ROUND_UP(offsetof(struct fw_vi_mac_cmd, t4vf_set_addr_hash() local 1783 size_t len16 = DIV_ROUND_UP(len, 16); t4vf_get_port_stats() local [all...] |
/kernel/linux/linux-6.6/fs/efivarfs/ |
H A D | vars.c | 343 unsigned long len16) in dup_variable_bug() 345 size_t i, len8 = len16 / sizeof(efi_char16_t); in dup_variable_bug() 342 dup_variable_bug(efi_char16_t *str16, efi_guid_t *vendor_guid, unsigned long len16) dup_variable_bug() argument
|
/kernel/linux/linux-5.10/drivers/firmware/efi/ |
H A D | vars.c | 385 unsigned long len16) in dup_variable_bug() 387 size_t i, len8 = len16 / sizeof(efi_char16_t); in dup_variable_bug() 384 dup_variable_bug(efi_char16_t *str16, efi_guid_t *vendor_guid, unsigned long len16) dup_variable_bug() argument
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls.h | 382 __be32 len16; /* command length */ member
|
H A D | chtls_hw.c | 411 kwr->req.len16 = cpu_to_be32((csk->tid << 8) | in chtls_setkey()
|
/kernel/linux/linux-5.10/drivers/target/iscsi/cxgbit/ |
H A D | cxgbit_ddp.c | 88 req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16)); in cxgbit_ppod_init_idata()
|
/kernel/linux/linux-6.6/drivers/target/iscsi/cxgbit/ |
H A D | cxgbit_ddp.c | 88 req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16)); in cxgbit_ppod_init_idata()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls.h | 385 __be32 len16; /* command length */ member
|
H A D | chtls_hw.c | 411 kwr->req.len16 = cpu_to_be32((csk->tid << 8) | in chtls_setkey()
|
/kernel/linux/linux-5.10/drivers/crypto/chelsio/ |
H A D | chcr_algo.c | 764 unsigned int len16, in create_wreq() 785 htonl(FW_CRYPTO_LOOKASIDE_WR_LEN16_V(DIV_ROUND_UP(len16, 16))); in create_wreq() 791 chcr_req->ulptx.len = htonl((DIV_ROUND_UP(len16, 16) - in create_wreq() 759 create_wreq(struct chcr_context *ctx, struct chcr_wr *chcr_req, struct crypto_async_request *req, unsigned int imm, int hash_sz, unsigned int len16, unsigned int sc_len, unsigned int lcb) create_wreq() argument
|