/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 124 u32 pg_size; member 294 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local 298 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size() 300 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() 303 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size() 306 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size() 309 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size() 312 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size() 315 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size() 321 return pg_size; in bnxt_qplib_base_pg_size() [all...] |
H A D | qplib_res.c | 72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 88 pbl->pg_size = 0; in __free_pbl() 132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl() 137 pbl->pg_size, in __alloc_pbl() 186 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local 196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq() 204 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq() 205 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq() 211 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq() 212 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq() [all...] |
H A D | qplib_sp.c | 662 u32 pg_size; in bnxt_qplib_reg_mr() local 706 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr() 711 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr() 713 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
|
/kernel/linux/linux-6.6/drivers/iommu/amd/ |
H A D | io_pgtable_v2.c | 63 static u64 set_pte_attr(u64 paddr, u64 pg_size, int prot) in set_pte_attr() argument 75 if (pg_size == IOMMU_PAGE_SIZE_1G || pg_size == IOMMU_PAGE_SIZE_2M) in set_pte_attr() 92 static inline int page_size_to_level(u64 pg_size) in page_size_to_level() argument 94 if (pg_size == IOMMU_PAGE_SIZE_1G) in page_size_to_level() 96 if (pg_size == IOMMU_PAGE_SIZE_2M) in page_size_to_level() 136 unsigned long pg_size, gfp_t gfp, bool *updated) in v2_alloc_pte() 142 end_level = page_size_to_level(pg_size); in v2_alloc_pte() 185 if (pg_size == IOMMU_PAGE_SIZE_1G) in v2_alloc_pte() 187 else if (pg_size in v2_alloc_pte() 135 v2_alloc_pte(int nid, u64 *pgd, unsigned long iova, unsigned long pg_size, gfp_t gfp, bool *updated) v2_alloc_pte() argument [all...] |
H A D | io_pgtable.c | 53 unsigned long pte_mask, pg_size, cnt; in first_pte_l7() local 56 pg_size = PTE_PAGE_SIZE(*pte); in first_pte_l7() 57 cnt = PAGE_SIZE_PTE_COUNT(pg_size); in first_pte_l7() 62 *page_size = pg_size; in first_pte_l7()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.h | 136 u32 pg_size; member 313 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local 317 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size() 319 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() 322 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size() 325 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size() 328 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size() 331 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size() 334 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size() 340 return pg_size; in bnxt_qplib_base_pg_size() [all...] |
H A D | qplib_res.c | 72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 88 pbl->pg_size = 0; in __free_pbl() 132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl() 137 pbl->pg_size, in __alloc_pbl() 186 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local 196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq() 204 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq() 205 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq() 211 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq() 212 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq() [all...] |
H A D | qplib_sp.c | 604 u32 pg_size; in bnxt_qplib_reg_mr() local 641 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr() 646 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr() 648 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
|
/kernel/linux/linux-6.6/fs/netfs/ |
H A D | iterator.c | 48 size_t bv_size, pg_size; in netfs_extract_user_iter() local 63 pg_size = array_size(max_pages, sizeof(*pages)); in netfs_extract_user_iter() 64 pages = (void *)bv + bv_size - pg_size; in netfs_extract_user_iter()
|
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/x86_64/ |
H A D | processor.c | 184 const uint64_t pg_size = PG_LEVEL_SIZE(level); in __virt_pg_map() local 191 TEST_ASSERT((vaddr % pg_size) == 0, in __virt_pg_map() 193 "vaddr: 0x%lx page size: 0x%lx", vaddr, pg_size); in __virt_pg_map() 196 TEST_ASSERT((paddr % pg_size) == 0, in __virt_pg_map() 198 " paddr: 0x%lx page size: 0x%lx", paddr, pg_size); in __virt_pg_map() 235 uint64_t pg_size = PG_LEVEL_SIZE(level); in virt_map_level() local 236 uint64_t nr_pages = nr_bytes / pg_size; in virt_map_level() 239 TEST_ASSERT(nr_bytes % pg_size == 0, in virt_map_level() 241 nr_bytes, pg_size); in virt_map_level() 246 vaddr += pg_size; in virt_map_level() [all...] |
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/rdma/ |
H A D | bnxt_re-abi.h | 13 __u32 pg_size; member
|
/kernel/linux/linux-5.10/include/uapi/rdma/ |
H A D | bnxt_re-abi.h | 58 __u32 pg_size; member
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/rdma/ |
H A D | bnxt_re-abi.h | 32 __u32 pg_size; member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/rdma/ |
H A D | bnxt_re-abi.h | 32 __u32 pg_size; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_eqs.c | 27 #define GET_EQ_NUM_PAGES(eq, pg_size) \ 28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) 30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size)
|
H A D | hinic_hw_io.c | 38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12)))
|
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_eqs.c | 27 #define GET_EQ_NUM_PAGES(eq, pg_size) \ 28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) 30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size)
|
H A D | hinic_hw_io.c | 38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12)))
|
/kernel/linux/linux-6.6/include/uapi/rdma/ |
H A D | bnxt_re-abi.h | 68 __u32 pg_size; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_io.c | 1106 int pg_size = PAGE_SIZE; in chtls_sendmsg() local 1111 pg_size = page_size(page); in chtls_sendmsg() 1112 if (off < pg_size && in chtls_sendmsg() 1122 if (page && off == pg_size) { in chtls_sendmsg() 1125 pg_size = PAGE_SIZE; in chtls_sendmsg() 1138 pg_size <<= order; in chtls_sendmsg() 1142 pg_size = PAGE_SIZE; in chtls_sendmsg() 1149 if (copy > pg_size - off) in chtls_sendmsg() 1150 copy = pg_size - off; in chtls_sendmsg() 1171 if (off + copy < pg_size) { in chtls_sendmsg() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
H A D | chtls_io.c | 1119 int pg_size = PAGE_SIZE; in chtls_sendmsg() local 1124 pg_size = page_size(page); in chtls_sendmsg() 1125 if (off < pg_size && in chtls_sendmsg() 1135 if (page && off == pg_size) { in chtls_sendmsg() 1138 pg_size = PAGE_SIZE; in chtls_sendmsg() 1151 pg_size <<= order; in chtls_sendmsg() 1155 pg_size = PAGE_SIZE; in chtls_sendmsg() 1162 if (copy > pg_size - off) in chtls_sendmsg() 1163 copy = pg_size - off; in chtls_sendmsg() 1184 if (off + copy < pg_size) { in chtls_sendmsg() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_verbs.c | 1322 * pg_size: page size 1325 static bool i40iw_check_mem_contiguous(u64 *arr, u32 npages, u32 pg_size) in i40iw_check_mem_contiguous() argument 1330 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in i40iw_check_mem_contiguous() 1339 * pg_size: page size 1341 static bool i40iw_check_mr_contiguous(struct i40iw_pble_alloc *palloc, u32 pg_size) in i40iw_check_mr_contiguous() argument 1352 ret = i40iw_check_mem_contiguous(arr, palloc->total_cnt, pg_size); in i40iw_check_mr_contiguous() 1360 if ((*start_addr + (i * pg_size * PBLE_PER_PAGE)) != *arr) in i40iw_check_mr_contiguous() 1362 ret = i40iw_check_mem_contiguous(arr, leaf->cnt, pg_size); in i40iw_check_mr_contiguous() 1428 u32 pg_size; in i40iw_handle_q_mem() local 1434 pg_size in i40iw_handle_q_mem() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/caam/ |
H A D | ctrl.c | 634 int pg_size; in caam_probe() local 704 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe() 705 if (pg_size == 0) in caam_probe()
|
/kernel/linux/linux-6.6/drivers/crypto/caam/ |
H A D | ctrl.c | 860 int pg_size; in caam_probe() local 974 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe() 975 if (pg_size == 0) in caam_probe()
|
/kernel/linux/linux-6.6/drivers/iommu/ |
H A D | iommu.c | 1057 unsigned long pg_size; in iommu_create_device_direct_mappings() local 1060 pg_size = domain->pgsize_bitmap ? 1UL << __ffs(domain->pgsize_bitmap) : 0; in iommu_create_device_direct_mappings() 1063 if (WARN_ON_ONCE(iommu_is_dma_domain(domain) && !pg_size)) in iommu_create_device_direct_mappings() 1081 start = ALIGN(entry->start, pg_size); in iommu_create_device_direct_mappings() 1082 end = ALIGN(entry->start + entry->length, pg_size); in iommu_create_device_direct_mappings() 1084 for (addr = start; addr <= end; addr += pg_size) { in iommu_create_device_direct_mappings() 1092 map_size += pg_size; in iommu_create_device_direct_mappings()
|