Home
last modified time | relevance | path

Searched refs:pg_size (Results 1 - 25 of 37) sorted by relevance

12

/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/
H A Dqplib_res.h124 u32 pg_size; member
294 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local
298 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size()
300 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size()
303 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size()
306 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size()
309 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size()
312 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size()
315 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size()
321 return pg_size; in bnxt_qplib_base_pg_size()
[all...]
H A Dqplib_res.c72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl()
88 pbl->pg_size = 0; in __free_pbl()
132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl()
137 pbl->pg_size, in __alloc_pbl()
186 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local
196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq()
204 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq()
205 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq()
211 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq()
212 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq()
[all...]
H A Dqplib_sp.c662 u32 pg_size; in bnxt_qplib_reg_mr() local
706 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr()
711 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr()
713 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
/kernel/linux/linux-6.6/drivers/iommu/amd/
H A Dio_pgtable_v2.c63 static u64 set_pte_attr(u64 paddr, u64 pg_size, int prot) in set_pte_attr() argument
75 if (pg_size == IOMMU_PAGE_SIZE_1G || pg_size == IOMMU_PAGE_SIZE_2M) in set_pte_attr()
92 static inline int page_size_to_level(u64 pg_size) in page_size_to_level() argument
94 if (pg_size == IOMMU_PAGE_SIZE_1G) in page_size_to_level()
96 if (pg_size == IOMMU_PAGE_SIZE_2M) in page_size_to_level()
136 unsigned long pg_size, gfp_t gfp, bool *updated) in v2_alloc_pte()
142 end_level = page_size_to_level(pg_size); in v2_alloc_pte()
185 if (pg_size == IOMMU_PAGE_SIZE_1G) in v2_alloc_pte()
187 else if (pg_size in v2_alloc_pte()
135 v2_alloc_pte(int nid, u64 *pgd, unsigned long iova, unsigned long pg_size, gfp_t gfp, bool *updated) v2_alloc_pte() argument
[all...]
H A Dio_pgtable.c53 unsigned long pte_mask, pg_size, cnt; in first_pte_l7() local
56 pg_size = PTE_PAGE_SIZE(*pte); in first_pte_l7()
57 cnt = PAGE_SIZE_PTE_COUNT(pg_size); in first_pte_l7()
62 *page_size = pg_size; in first_pte_l7()
/kernel/linux/linux-6.6/drivers/infiniband/hw/bnxt_re/
H A Dqplib_res.h136 u32 pg_size; member
313 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local
317 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size()
319 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size()
322 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size()
325 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size()
328 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size()
331 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size()
334 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size()
340 return pg_size; in bnxt_qplib_base_pg_size()
[all...]
H A Dqplib_res.c72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl()
88 pbl->pg_size = 0; in __free_pbl()
132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl()
137 pbl->pg_size, in __alloc_pbl()
186 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local
196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq()
204 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq()
205 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq()
211 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq()
212 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq()
[all...]
H A Dqplib_sp.c604 u32 pg_size; in bnxt_qplib_reg_mr() local
641 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr()
646 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr()
648 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
/kernel/linux/linux-6.6/fs/netfs/
H A Diterator.c48 size_t bv_size, pg_size; in netfs_extract_user_iter() local
63 pg_size = array_size(max_pages, sizeof(*pages)); in netfs_extract_user_iter()
64 pages = (void *)bv + bv_size - pg_size; in netfs_extract_user_iter()
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/x86_64/
H A Dprocessor.c184 const uint64_t pg_size = PG_LEVEL_SIZE(level); in __virt_pg_map() local
191 TEST_ASSERT((vaddr % pg_size) == 0, in __virt_pg_map()
193 "vaddr: 0x%lx page size: 0x%lx", vaddr, pg_size); in __virt_pg_map()
196 TEST_ASSERT((paddr % pg_size) == 0, in __virt_pg_map()
198 " paddr: 0x%lx page size: 0x%lx", paddr, pg_size); in __virt_pg_map()
235 uint64_t pg_size = PG_LEVEL_SIZE(level); in virt_map_level() local
236 uint64_t nr_pages = nr_bytes / pg_size; in virt_map_level()
239 TEST_ASSERT(nr_bytes % pg_size == 0, in virt_map_level()
241 nr_bytes, pg_size); in virt_map_level()
246 vaddr += pg_size; in virt_map_level()
[all...]
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/rdma/
H A Dbnxt_re-abi.h13 __u32 pg_size; member
/kernel/linux/linux-5.10/include/uapi/rdma/
H A Dbnxt_re-abi.h58 __u32 pg_size; member
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/rdma/
H A Dbnxt_re-abi.h32 __u32 pg_size; member
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/rdma/
H A Dbnxt_re-abi.h32 __u32 pg_size; member
/kernel/linux/linux-5.10/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_eqs.c27 #define GET_EQ_NUM_PAGES(eq, pg_size) \
28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size))
30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size)
H A Dhinic_hw_io.c38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12)))
/kernel/linux/linux-6.6/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_eqs.c27 #define GET_EQ_NUM_PAGES(eq, pg_size) \
28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size))
30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size)
H A Dhinic_hw_io.c38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12)))
/kernel/linux/linux-6.6/include/uapi/rdma/
H A Dbnxt_re-abi.h68 __u32 pg_size; member
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_io.c1106 int pg_size = PAGE_SIZE; in chtls_sendmsg() local
1111 pg_size = page_size(page); in chtls_sendmsg()
1112 if (off < pg_size && in chtls_sendmsg()
1122 if (page && off == pg_size) { in chtls_sendmsg()
1125 pg_size = PAGE_SIZE; in chtls_sendmsg()
1138 pg_size <<= order; in chtls_sendmsg()
1142 pg_size = PAGE_SIZE; in chtls_sendmsg()
1149 if (copy > pg_size - off) in chtls_sendmsg()
1150 copy = pg_size - off; in chtls_sendmsg()
1171 if (off + copy < pg_size) { in chtls_sendmsg()
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/inline_crypto/chtls/
H A Dchtls_io.c1119 int pg_size = PAGE_SIZE; in chtls_sendmsg() local
1124 pg_size = page_size(page); in chtls_sendmsg()
1125 if (off < pg_size && in chtls_sendmsg()
1135 if (page && off == pg_size) { in chtls_sendmsg()
1138 pg_size = PAGE_SIZE; in chtls_sendmsg()
1151 pg_size <<= order; in chtls_sendmsg()
1155 pg_size = PAGE_SIZE; in chtls_sendmsg()
1162 if (copy > pg_size - off) in chtls_sendmsg()
1163 copy = pg_size - off; in chtls_sendmsg()
1184 if (off + copy < pg_size) { in chtls_sendmsg()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/
H A Di40iw_verbs.c1322 * pg_size: page size
1325 static bool i40iw_check_mem_contiguous(u64 *arr, u32 npages, u32 pg_size) in i40iw_check_mem_contiguous() argument
1330 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in i40iw_check_mem_contiguous()
1339 * pg_size: page size
1341 static bool i40iw_check_mr_contiguous(struct i40iw_pble_alloc *palloc, u32 pg_size) in i40iw_check_mr_contiguous() argument
1352 ret = i40iw_check_mem_contiguous(arr, palloc->total_cnt, pg_size); in i40iw_check_mr_contiguous()
1360 if ((*start_addr + (i * pg_size * PBLE_PER_PAGE)) != *arr) in i40iw_check_mr_contiguous()
1362 ret = i40iw_check_mem_contiguous(arr, leaf->cnt, pg_size); in i40iw_check_mr_contiguous()
1428 u32 pg_size; in i40iw_handle_q_mem() local
1434 pg_size in i40iw_handle_q_mem()
[all...]
/kernel/linux/linux-5.10/drivers/crypto/caam/
H A Dctrl.c634 int pg_size; in caam_probe() local
704 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe()
705 if (pg_size == 0) in caam_probe()
/kernel/linux/linux-6.6/drivers/crypto/caam/
H A Dctrl.c860 int pg_size; in caam_probe() local
974 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe()
975 if (pg_size == 0) in caam_probe()
/kernel/linux/linux-6.6/drivers/iommu/
H A Diommu.c1057 unsigned long pg_size; in iommu_create_device_direct_mappings() local
1060 pg_size = domain->pgsize_bitmap ? 1UL << __ffs(domain->pgsize_bitmap) : 0; in iommu_create_device_direct_mappings()
1063 if (WARN_ON_ONCE(iommu_is_dma_domain(domain) && !pg_size)) in iommu_create_device_direct_mappings()
1081 start = ALIGN(entry->start, pg_size); in iommu_create_device_direct_mappings()
1082 end = ALIGN(entry->start + entry->length, pg_size); in iommu_create_device_direct_mappings()
1084 for (addr = start; addr <= end; addr += pg_size) { in iommu_create_device_direct_mappings()
1092 map_size += pg_size; in iommu_create_device_direct_mappings()

Completed in 27 milliseconds

12