Home
last modified time | relevance | path

Searched refs:chunk_list (Results 1 - 25 of 36) sorted by relevance

12

/kernel/linux/linux-5.10/arch/powerpc/platforms/ps3/
H A Dmm.c364 * @link: A struct list_head used with struct ps3_dma_region.chunk_list, the
404 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
433 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
539 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
573 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
578 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
609 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
659 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
660 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
681 INIT_LIST_HEAD(&r->chunk_list in dma_ioc0_region_create()
[all...]
/kernel/linux/linux-6.6/arch/powerpc/platforms/ps3/
H A Dmm.c365 * @link: A struct list_head used with struct ps3_dma_region.chunk_list, the
405 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
434 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
540 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
574 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
579 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
610 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
660 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
661 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
682 INIT_LIST_HEAD(&r->chunk_list in dma_ioc0_region_create()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/
H A Dusnic_uiom.c66 static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) in usnic_uiom_put_pages() argument
74 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
88 struct list_head *chunk_list = &uiomr->chunk_list; in usnic_uiom_get_pages() local
117 INIT_LIST_HEAD(chunk_list); in usnic_uiom_get_pages()
177 list_add_tail(&chunk->list, chunk_list); in usnic_uiom_get_pages()
185 usnic_uiom_put_pages(chunk_list, 0); in usnic_uiom_get_pages()
240 usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable); in __usnic_uiom_reg_release()
260 chunk = list_first_entry(&uiomr->chunk_list, struct usnic_uiom_chunk, in usnic_uiom_map_sorted_intervals()
415 usnic_uiom_put_pages(&uiomr->chunk_list, in usnic_uiom_reg_get()
[all...]
H A Dusnic_uiom.h72 struct list_head chunk_list; member
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_icm_pool.c258 INIT_LIST_HEAD(&chunk->chunk_list); in dr_icm_chunks_create()
259 list_add(&chunk->chunk_list, &bucket->free_list); in dr_icm_chunks_create()
284 list_del(&chunk->chunk_list); in dr_icm_chunk_destroy()
319 list_for_each_entry_safe(chunk, next, &bucket->free_list, chunk_list) in dr_icm_bucket_cleanup()
325 list_for_each_entry_safe(chunk, next, &bucket->used_list, chunk_list) in dr_icm_bucket_cleanup()
481 chunk_list); in mlx5dr_icm_alloc_chunk()
483 list_del_init(&chunk->chunk_list); in mlx5dr_icm_alloc_chunk()
484 list_add_tail(&chunk->chunk_list, &bucket->used_list); in mlx5dr_icm_alloc_chunk()
506 list_del_init(&chunk->chunk_list); in mlx5dr_icm_free_chunk()
507 list_add_tail(&chunk->chunk_list, in mlx5dr_icm_free_chunk()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/usnic/
H A Dusnic_uiom.c65 static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) in usnic_uiom_put_pages() argument
73 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
87 struct list_head *chunk_list = &uiomr->chunk_list; in usnic_uiom_get_pages() local
115 INIT_LIST_HEAD(chunk_list); in usnic_uiom_get_pages()
172 list_add_tail(&chunk->list, chunk_list); in usnic_uiom_get_pages()
180 usnic_uiom_put_pages(chunk_list, 0); in usnic_uiom_get_pages()
235 usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable); in __usnic_uiom_reg_release()
255 chunk = list_first_entry(&uiomr->chunk_list, struct usnic_uiom_chunk, in usnic_uiom_map_sorted_intervals()
410 usnic_uiom_put_pages(&uiomr->chunk_list, in usnic_uiom_reg_get()
[all...]
H A Dusnic_uiom.h72 struct list_head chunk_list; member
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
H A Dhns_roce_hem.h83 struct list_head chunk_list; member
140 iter->chunk = list_empty(&hem->chunk_list) ? NULL : in hns_roce_hem_first()
141 list_entry(hem->chunk_list.next, in hns_roce_hem_first()
154 if (iter->chunk->list.next == &iter->hem->chunk_list) { in hns_roce_hem_next()
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/
H A Dicm.h68 struct list_head chunk_list; member
100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
101 NULL : list_entry(icm->chunk_list.next, in mlx4_icm_first()
114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
H A Dicm.c86 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mlx4_free_icm()
155 INIT_LIST_HEAD(&icm->chunk_list); in mlx4_alloc_icm()
176 list_add_tail(&chunk->list, &icm->chunk_list); in mlx4_alloc_icm()
335 list_for_each_entry(chunk, &icm->chunk_list, list) { in mlx4_table_find()
/kernel/linux/linux-6.6/drivers/infiniband/hw/hns/
H A Dhns_roce_hem.h84 struct list_head chunk_list; member
140 iter->chunk = list_empty(&hem->chunk_list) ? NULL : in hns_roce_hem_first()
141 list_entry(hem->chunk_list.next, in hns_roce_hem_first()
154 if (iter->chunk->list.next == &iter->hem->chunk_list) { in hns_roce_hem_next()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx4/
H A Dicm.h68 struct list_head chunk_list; member
100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
101 NULL : list_entry(icm->chunk_list.next, in mlx4_icm_first()
114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
H A Dicm.c86 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mlx4_free_icm()
155 INIT_LIST_HEAD(&icm->chunk_list); in mlx4_alloc_icm()
176 list_add_tail(&chunk->list, &icm->chunk_list); in mlx4_alloc_icm()
335 list_for_each_entry(chunk, &icm->chunk_list, list) { in mlx4_table_find()
/kernel/linux/linux-5.10/drivers/infiniband/hw/efa/
H A Defa_verbs.c105 struct pbl_chunk_list chunk_list; member
1191 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() local
1205 chunk_list->size = chunk_list_size; in pbl_chunk_list_create()
1206 chunk_list->chunks = kcalloc(chunk_list_size, in pbl_chunk_list_create()
1207 sizeof(*chunk_list->chunks), in pbl_chunk_list_create()
1209 if (!chunk_list->chunks) in pbl_chunk_list_create()
1218 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create()
1219 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create()
1222 chunk_list in pbl_chunk_list_create()
1291 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; pbl_chunk_list_destroy() local
[all...]
/kernel/linux/linux-5.10/net/sctp/
H A Doutput.c151 INIT_LIST_HEAD(&packet->chunk_list); in sctp_packet_init()
165 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
347 list_add_tail(&chunk->list, &packet->chunk_list); in __sctp_packet_append_chunk()
429 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, in sctp_packet_pack()
450 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_pack()
493 if (list_empty(&packet->chunk_list)) in sctp_packet_pack()
497 &packet->chunk_list); in sctp_packet_pack()
504 } while (!list_empty(&packet->chunk_list)); in sctp_packet_pack()
557 if (list_empty(&packet->chunk_list)) in sctp_packet_transmit()
559 chunk = list_entry(packet->chunk_list in sctp_packet_transmit()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
H A Dmthca_memfree.h59 struct list_head chunk_list; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
104 NULL : list_entry(icm->chunk_list.next, in mthca_icm_first()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
H A Dmthca_memfree.c95 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mthca_free_icm()
153 INIT_LIST_HEAD(&icm->chunk_list); in mthca_alloc_icm()
167 list_add_tail(&chunk->list, &icm->chunk_list); in mthca_alloc_icm()
297 list_for_each_entry(chunk, &icm->chunk_list, list) { in mthca_table_find()
/kernel/linux/linux-6.6/drivers/infiniband/hw/mthca/
H A Dmthca_memfree.h59 struct list_head chunk_list; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
104 NULL : list_entry(icm->chunk_list.next, in mthca_icm_first()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
H A Dmthca_memfree.c95 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mthca_free_icm()
153 INIT_LIST_HEAD(&icm->chunk_list); in mthca_alloc_icm()
167 list_add_tail(&chunk->list, &icm->chunk_list); in mthca_alloc_icm()
297 list_for_each_entry(chunk, &icm->chunk_list, list) { in mthca_table_find()
/kernel/linux/linux-6.6/net/sctp/
H A Doutput.c153 INIT_LIST_HEAD(&packet->chunk_list); in sctp_packet_init()
167 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
233 list_add_tail(&pad->list, &pkt->chunk_list); in sctp_packet_bundle_pad()
373 list_add_tail(&chunk->list, &packet->chunk_list); in __sctp_packet_append_chunk()
459 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, in sctp_packet_pack()
480 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_pack()
523 if (list_empty(&packet->chunk_list)) in sctp_packet_pack()
527 &packet->chunk_list); in sctp_packet_pack()
534 } while (!list_empty(&packet->chunk_list)); in sctp_packet_pack()
580 if (list_empty(&packet->chunk_list)) in sctp_packet_transmit()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/efa/
H A Defa_verbs.c123 struct pbl_chunk_list chunk_list; member
1270 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() local
1284 chunk_list->size = chunk_list_size; in pbl_chunk_list_create()
1285 chunk_list->chunks = kcalloc(chunk_list_size, in pbl_chunk_list_create()
1286 sizeof(*chunk_list->chunks), in pbl_chunk_list_create()
1288 if (!chunk_list->chunks) in pbl_chunk_list_create()
1297 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create()
1298 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create()
1301 chunk_list in pbl_chunk_list_create()
1370 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; pbl_chunk_list_destroy() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/qxl/
H A Dqxl_image.c51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
68 INIT_LIST_HEAD(&image->chunk_list); in qxl_image_alloc_objects()
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
120 drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); in qxl_image_init_helper()
/kernel/linux/linux-6.6/drivers/gpu/drm/qxl/
H A Dqxl_image.c51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
68 INIT_LIST_HEAD(&image->chunk_list); in qxl_image_alloc_objects()
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
120 drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); in qxl_image_init_helper()
/kernel/linux/linux-5.10/drivers/s390/block/
H A Ddasd_int.h650 dasd_init_chunklist(struct list_head *chunk_list, void *mem, in dasd_init_chunklist() argument
655 INIT_LIST_HEAD(chunk_list); in dasd_init_chunklist()
658 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
662 dasd_alloc_chunk(struct list_head *chunk_list, unsigned long size) in dasd_alloc_chunk() argument
667 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
684 dasd_free_chunk(struct list_head *chunk_list, void *mem) in dasd_free_chunk() argument
691 /* Find out the left neighbour in chunk_list. */ in dasd_free_chunk()
692 left = chunk_list; in dasd_free_chunk()
693 list_for_each(p, chunk_list) { in dasd_free_chunk()
699 if (left->next != chunk_list) { in dasd_free_chunk()
[all...]
/kernel/linux/linux-6.6/drivers/s390/block/
H A Ddasd_int.h759 dasd_init_chunklist(struct list_head *chunk_list, void *mem, in dasd_init_chunklist() argument
764 INIT_LIST_HEAD(chunk_list); in dasd_init_chunklist()
767 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
771 dasd_alloc_chunk(struct list_head *chunk_list, unsigned long size) in dasd_alloc_chunk() argument
776 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
793 dasd_free_chunk(struct list_head *chunk_list, void *mem) in dasd_free_chunk() argument
800 /* Find out the left neighbour in chunk_list. */ in dasd_free_chunk()
801 left = chunk_list; in dasd_free_chunk()
802 list_for_each(p, chunk_list) { in dasd_free_chunk()
808 if (left->next != chunk_list) { in dasd_free_chunk()
[all...]

Completed in 22 milliseconds

12