/kernel/linux/linux-5.10/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_chain.c | 12 u32 page_cnt) in qed_chain_init() 34 chain->page_cnt = page_cnt; in qed_chain_init() 35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init() 36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init() 80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr() 114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl() 160 u32 page_cnt) in qed_chain_alloc_sanity_check() 165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check() 206 for (i = 0; i < chain->page_cnt; in qed_chain_alloc_next_ptr() 10 qed_chain_init(struct qed_chain *chain, const struct qed_chain_init_params *params, u32 page_cnt) qed_chain_init() argument 158 qed_chain_alloc_sanity_check(struct qed_dev *cdev, const struct qed_chain_init_params *params, u32 page_cnt) qed_chain_alloc_sanity_check() argument 255 u32 page_cnt, i; qed_chain_alloc_pbl() local 323 u32 page_cnt; qed_chain_alloc() local [all...] |
H A D | qed_sp_commands.c | 309 u8 page_cnt, i; in qed_sp_pf_start() local 370 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start() 371 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_chain.c | 12 u32 page_cnt) in qed_chain_init() 34 chain->page_cnt = page_cnt; in qed_chain_init() 35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init() 36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init() 80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr() 114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl() 160 u32 page_cnt) in qed_chain_alloc_sanity_check() 165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check() 206 for (i = 0; i < chain->page_cnt; in qed_chain_alloc_next_ptr() 10 qed_chain_init(struct qed_chain *chain, const struct qed_chain_init_params *params, u32 page_cnt) qed_chain_init() argument 158 qed_chain_alloc_sanity_check(struct qed_dev *cdev, const struct qed_chain_init_params *params, u32 page_cnt) qed_chain_alloc_sanity_check() argument 255 u32 page_cnt, i; qed_chain_alloc_pbl() local 323 u32 page_cnt; qed_chain_alloc() local [all...] |
H A D | qed_sp_commands.c | 311 u8 page_cnt, i; in qed_sp_pf_start() local 372 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start() 373 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start() 378 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_consq->chain); in qed_sp_pf_start() 379 p_ramrod->consolid_q_num_pages = page_cnt; in qed_sp_pf_start()
|
/kernel/linux/linux-5.10/include/linux/qed/ |
H A D | qed_chain.h | 106 u32 page_cnt; member 262 return chain->page_cnt; in qed_chain_get_page_cnt() 302 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 306 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 490 /* Use (page_cnt - 1) as a reset value for the prod/cons page's in qed_chain_reset() 493 * will be advanced to page_cnt and then will be wrapped to 0. in qed_chain_reset() 495 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset() 554 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem() 575 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local 594 page_cnt in qed_chain_set_prod() 619 u32 i, page_cnt; qed_chain_pbl_zero_mem() local [all...] |
/kernel/linux/linux-6.6/include/linux/qed/ |
H A D | qed_chain.h | 106 u32 page_cnt; member 262 return chain->page_cnt; in qed_chain_get_page_cnt() 303 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 307 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 493 /* Use (page_cnt - 1) as a reset value for the prod/cons page's in qed_chain_reset() 496 * will be advanced to page_cnt and then will be wrapped to 0. in qed_chain_reset() 498 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset() 556 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem() 580 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local 599 page_cnt in qed_chain_set_prod() 626 u32 i, page_cnt; qed_chain_pbl_zero_mem() local [all...] |
/kernel/linux/linux-5.10/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-helper-util.c | 133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local 141 page_cnt.u64 = 0; in cvmx_helper_setup_red() 142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red() 143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red() 148 page_cnt.u64); in cvmx_helper_setup_red()
|
/kernel/linux/linux-6.6/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-helper-util.c | 133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local 141 page_cnt.u64 = 0; in cvmx_helper_setup_red() 142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red() 143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red() 148 page_cnt.u64); in cvmx_helper_setup_red()
|
/kernel/linux/linux-5.10/drivers/hv/ |
H A D | ring_buffer.c | 193 struct page *pages, u32 page_cnt) in hv_ringbuffer_init() 204 pages_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(struct page *), in hv_ringbuffer_init() 210 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init() 211 pages_wraparound[i + 1] = &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init() 214 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, PAGE_KERNEL); in hv_ringbuffer_init() 228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init() 192 hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info, struct page *pages, u32 page_cnt) hv_ringbuffer_init() argument
|
H A D | hv_balloon.c | 142 __u64 page_cnt:24; member 994 pfn_cnt = dm->ha_wrk.ha_page_range.finfo.page_cnt; in hot_add_req() 997 rg_sz = dm->ha_wrk.ha_region_range.finfo.page_cnt; in hot_add_req() 1191 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages() 1247 bl_resp->range_array[i].finfo.page_cnt = alloc_unit; in alloc_balloon_pages()
|
/kernel/linux/linux-6.6/drivers/base/firmware_loader/ |
H A D | sysfs.c | 246 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local 250 page_ofs, page_cnt); in firmware_rw() 253 buffer, page_cnt); in firmware_rw() 255 buffer += page_cnt; in firmware_rw() 256 offset += page_cnt; in firmware_rw() 257 count -= page_cnt; in firmware_rw()
|
/kernel/linux/linux-5.10/drivers/base/firmware_loader/ |
H A D | fallback.c | 321 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local 326 memcpy(buffer, page_data + page_ofs, page_cnt); in firmware_rw() 328 memcpy(page_data + page_ofs, buffer, page_cnt); in firmware_rw() 331 buffer += page_cnt; in firmware_rw() 332 offset += page_cnt; in firmware_rw() 333 count -= page_cnt; in firmware_rw()
|
/kernel/linux/linux-6.6/drivers/hv/ |
H A D | ring_buffer.c | 186 struct page *pages, u32 page_cnt, u32 max_pkt_size) in hv_ringbuffer_init() 197 pages_wraparound = kcalloc(page_cnt * 2 - 1, in hv_ringbuffer_init() 204 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init() 206 &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init() 209 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, in hv_ringbuffer_init() 228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init() 185 hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info, struct page *pages, u32 page_cnt, u32 max_pkt_size) hv_ringbuffer_init() argument
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 843 dma_addr_t *pages, int page_cnt) in hns_roce_mtr_map() 861 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map() 865 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map() 947 int page_cnt = 0; in mtr_init_buf_cfg() local 953 page_cnt = DIV_ROUND_UP(buf_size, HNS_HW_PAGE_SIZE); in mtr_init_buf_cfg() 958 cfg->buf_pg_shift = HNS_HW_PAGE_SHIFT + order_base_2(page_cnt); in mtr_init_buf_cfg() 960 cfg->buf_pg_count = page_cnt; in mtr_init_buf_cfg() 977 for (page_cnt = 0, region_cnt = 0; page_cnt < cfg->buf_pg_count && in mtr_init_buf_cfg() 981 r->offset = page_cnt; in mtr_init_buf_cfg() 842 hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr, dma_addr_t *pages, int page_cnt) hns_roce_mtr_map() argument [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 746 dma_addr_t *pages, unsigned int page_cnt) in hns_roce_mtr_map() 763 mapped_cnt < page_cnt; i++) { in hns_roce_mtr_map() 771 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map() 775 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map() 780 page_cnt - mapped_cnt); in hns_roce_mtr_map() 791 if (mapped_cnt < page_cnt) { in hns_roce_mtr_map() 794 mapped_cnt, page_cnt); in hns_roce_mtr_map() 862 int page_cnt, region_cnt; in mtr_init_buf_cfg() local 894 for (page_cnt = 0, region_cnt = 0; region_cnt < attr->region_count && in mtr_init_buf_cfg() 897 r->offset = page_cnt; in mtr_init_buf_cfg() 745 hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr, dma_addr_t *pages, unsigned int page_cnt) hns_roce_mtr_map() argument [all...] |
/kernel/linux/linux-5.10/arch/mips/include/asm/octeon/ |
H A D | cvmx-ipd-defs.h | 324 uint64_t page_cnt:17; member 326 uint64_t page_cnt:17; 999 uint64_t page_cnt:17; member 1001 uint64_t page_cnt:17; 1014 uint64_t page_cnt:17; member 1016 uint64_t page_cnt:17; 1029 uint64_t page_cnt:17; member 1031 uint64_t page_cnt:17; 1387 uint64_t page_cnt:25; member 1389 uint64_t page_cnt [all...] |
/kernel/linux/linux-6.6/arch/mips/include/asm/octeon/ |
H A D | cvmx-ipd-defs.h | 324 uint64_t page_cnt:17; member 326 uint64_t page_cnt:17; 999 uint64_t page_cnt:17; member 1001 uint64_t page_cnt:17; 1014 uint64_t page_cnt:17; member 1016 uint64_t page_cnt:17; 1029 uint64_t page_cnt:17; member 1031 uint64_t page_cnt:17; 1387 uint64_t page_cnt:25; member 1389 uint64_t page_cnt [all...] |
/kernel/linux/linux-5.10/drivers/staging/rts5208/ |
H A D | xd.c | 1506 u8 reg_val, page_cnt; in xd_read_multiple_pages() local 1512 page_cnt = end_page - start_page; in xd_read_multiple_pages() 1538 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_read_multiple_pages() 1543 page_cnt * 512, DMA_512); in xd_read_multiple_pages() 1553 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_read_multiple_pages() 1709 u8 page_cnt, reg_val; in xd_write_multiple_pages() local 1717 page_cnt = end_page - start_page; in xd_write_multiple_pages() 1739 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_write_multiple_pages() 1743 page_cnt * 512, DMA_512); in xd_write_multiple_pages() 1752 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 51 in xd_write_multiple_pages() 1840 u8 start_page, end_page = 0, page_cnt; xd_rw() local [all...] |
/kernel/linux/linux-6.6/drivers/staging/rts5208/ |
H A D | xd.c | 1499 u8 reg_val, page_cnt; in xd_read_multiple_pages() local 1505 page_cnt = end_page - start_page; in xd_read_multiple_pages() 1531 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_read_multiple_pages() 1536 page_cnt * 512, DMA_512); in xd_read_multiple_pages() 1546 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_read_multiple_pages() 1701 u8 page_cnt, reg_val; in xd_write_multiple_pages() local 1709 page_cnt = end_page - start_page; in xd_write_multiple_pages() 1731 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_write_multiple_pages() 1735 page_cnt * 512, DMA_512); in xd_write_multiple_pages() 1744 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 51 in xd_write_multiple_pages() 1831 u8 start_page, end_page = 0, page_cnt; xd_rw() local [all...] |
/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | fslog.c | 1584 u32 page_cnt = 1, page_pos = 1; in last_log_lsn() local 1707 page_cnt = le16_to_cpu(best_page->page_count); in last_log_lsn() 1708 if (page_cnt > 1) in last_log_lsn() 1716 (page_pos != page_cnt || best_page_pos == page_pos || in last_log_lsn() 1718 (page_pos >= page_cnt || best_page_pos == page_pos)) { in last_log_lsn() 1730 page_cnt = bppc; in last_log_lsn() 1733 if (page_cnt > 1) in last_log_lsn() 1736 page_pos = page_cnt = 1; in last_log_lsn() 1766 page_cnt = page_pos = 1; in last_log_lsn() 1808 ((page_pos == page_cnt in last_log_lsn() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
H A D | i40iw_verbs.h | 95 u32 page_cnt; member
|
/kernel/linux/linux-6.6/arch/x86/kernel/cpu/sgx/ |
H A D | encl.h | 55 unsigned int page_cnt; member
|
H A D | ioctl.c | 28 if (!(encl->page_cnt % SGX_VA_SLOT_COUNT)) { in sgx_encl_grow() 40 WARN_ON_ONCE(encl->page_cnt % SGX_VA_SLOT_COUNT); in sgx_encl_grow() 42 encl->page_cnt++; in sgx_encl_grow() 48 encl->page_cnt--; in sgx_encl_shrink()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 1161 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument 1167 sglist = kmalloc_array(page_cnt, sizeof(*sglist), GFP_KERNEL); in efa_vmalloc_buf_to_sg() 1170 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg() 1171 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg() 1192 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local 1203 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create() 1214 page_cnt); in pbl_chunk_list_create() 1225 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 142 mr->mem.page_cnt > ERDMA_MAX_INLINE_MTT_ENTRIES) { in regmr_cmd() 152 MTT_SIZE(mr->mem.page_cnt)); in regmr_cmd() 164 FIELD_PREP(ERDMA_CMD_REGMR_MTT_CNT_MASK, mr->mem.page_cnt); in regmr_cmd() 179 req.mtt_cnt_h = mr->mem.page_cnt >> 20; in regmr_cmd() 772 mem->page_cnt = mem->mtt_nents; in get_mtt_entries() 773 mem->mtt = erdma_create_mtt(dev, MTT_SIZE(mem->page_cnt), in get_mtt_entries() 1091 mr->mem.page_cnt = max_num_sg; in erdma_ib_alloc_mr() 1121 if (mr->mem.mtt_nents >= mr->mem.page_cnt) in erdma_set_page()
|