Searched refs:dma_pages (Results 1 - 7 of 7) sorted by relevance
/kernel/linux/linux-5.10/drivers/gpu/drm/savage/ |
H A D | savage_bci.c | 310 dev_priv->dma_pages = kmalloc_array(dev_priv->nr_dma_pages, in savage_dma_init() 313 if (dev_priv->dma_pages == NULL) in savage_dma_init() 317 SET_AGE(&dev_priv->dma_pages[i].age, 0, 0); in savage_dma_init() 318 dev_priv->dma_pages[i].used = 0; in savage_dma_init() 319 dev_priv->dma_pages[i].flushed = 0; in savage_dma_init() 336 SET_AGE(&dev_priv->dma_pages[i].age, event, wrap); in savage_dma_reset() 337 dev_priv->dma_pages[i].used = 0; in savage_dma_reset() 338 dev_priv->dma_pages[i].flushed = 0; in savage_dma_reset() 362 if (dev_priv->dma_pages[page].age.wrap > wrap || in savage_dma_wait() 363 (dev_priv->dma_pages[pag in savage_dma_wait() [all...] |
H A D | savage_drv.h | 178 drm_savage_dma_page_t *dma_pages; member 520 dev_priv->dma_pages[cur].used; \ 526 dev_priv->dma_pages[cur].used; \ 527 if (dev_priv->dma_pages[cur].used == 0) \ 529 dev_priv->dma_pages[cur].used += (n); \ 545 dev_priv->dma_pages[cur].used; \
|
/kernel/linux/linux-6.6/net/xdp/ |
H A D | xsk_buff_pool.c | 208 if (!pool->dma_pages) { in xp_assign_dev() 319 dma_map->dma_pages = kvcalloc(nr_pages, sizeof(*dma_map->dma_pages), GFP_KERNEL); in xp_create_dma_map() 320 if (!dma_map->dma_pages) { in xp_create_dma_map() 337 kvfree(dma_map->dma_pages); in xp_destroy_dma_map() 347 dma = &dma_map->dma_pages[i]; in __xp_dma_unmap() 363 if (!pool->dma_pages) in xp_dma_unmap() 376 kvfree(pool->dma_pages); in xp_dma_unmap() 377 pool->dma_pages = NULL; in xp_dma_unmap() 388 if (dma_map->dma_pages[ in xp_check_dma_contiguity() [all...] |
/kernel/linux/linux-5.10/net/xdp/ |
H A D | xsk_buff_pool.c | 175 if (!pool->dma_pages) { in __xp_assign_dev() 292 dma_map->dma_pages = kvcalloc(nr_pages, sizeof(*dma_map->dma_pages), GFP_KERNEL); in xp_create_dma_map() 293 if (!dma_map->dma_pages) { in xp_create_dma_map() 310 kvfree(dma_map->dma_pages); in xp_destroy_dma_map() 320 dma = &dma_map->dma_pages[i]; in __xp_dma_unmap() 349 kvfree(pool->dma_pages); in xp_dma_unmap() 360 if (dma_map->dma_pages[i] + PAGE_SIZE == dma_map->dma_pages[i + 1]) in xp_check_dma_contiguity() 361 dma_map->dma_pages[ in xp_check_dma_contiguity() [all...] |
/kernel/linux/linux-5.10/arch/c6x/mm/ |
H A D | dma-coherent.c | 33 static u32 dma_pages; variable 49 pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); in __alloc_dma_pages() 60 if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { in __free_dma_pages() 131 dma_pages = dma_size >> PAGE_SHIFT; in coherent_mem_init() 133 ++dma_pages; in coherent_mem_init() 135 dma_bitmap = memblock_alloc(BITS_TO_LONGS(dma_pages) * sizeof(long), in coherent_mem_init() 139 __func__, BITS_TO_LONGS(dma_pages) * sizeof(long), in coherent_mem_init()
|
/kernel/linux/linux-6.6/include/net/ |
H A D | xsk_buff_pool.h | 38 dma_addr_t *dma_pages; member 65 /* For performance reasons, each buff pool has its own array of dma_pages 68 dma_addr_t *dma_pages; member 126 dma_addr_t *dma_pages, u64 addr) in xp_init_xskb_dma() 128 xskb->frame_dma = (dma_pages[addr >> PAGE_SHIFT] & ~XSK_NEXT_PG_CONTIG_MASK) + in xp_init_xskb_dma() 185 return pool->dma_pages && in xp_desc_crosses_non_contig_pg() 186 !(pool->dma_pages[addr >> PAGE_SHIFT] & XSK_NEXT_PG_CONTIG_MASK); in xp_desc_crosses_non_contig_pg() 125 xp_init_xskb_dma(struct xdp_buff_xsk *xskb, struct xsk_buff_pool *pool, dma_addr_t *dma_pages, u64 addr) xp_init_xskb_dma() argument
|
/kernel/linux/linux-5.10/include/net/ |
H A D | xsk_buff_pool.h | 32 dma_addr_t *dma_pages; member 58 /* For performance reasons, each buff pool has its own array of dma_pages 61 dma_addr_t *dma_pages; member 154 !(pool->dma_pages[addr >> PAGE_SHIFT] & XSK_NEXT_PG_CONTIG_MASK); in xp_desc_crosses_non_contig_pg()
|
Completed in 6 milliseconds