Home
last modified time | relevance | path

Searched refs:dma_pages (Results 1 - 7 of 7) sorted by relevance

/kernel/linux/linux-5.10/drivers/gpu/drm/savage/
H A Dsavage_bci.c310 dev_priv->dma_pages = kmalloc_array(dev_priv->nr_dma_pages, in savage_dma_init()
313 if (dev_priv->dma_pages == NULL) in savage_dma_init()
317 SET_AGE(&dev_priv->dma_pages[i].age, 0, 0); in savage_dma_init()
318 dev_priv->dma_pages[i].used = 0; in savage_dma_init()
319 dev_priv->dma_pages[i].flushed = 0; in savage_dma_init()
336 SET_AGE(&dev_priv->dma_pages[i].age, event, wrap); in savage_dma_reset()
337 dev_priv->dma_pages[i].used = 0; in savage_dma_reset()
338 dev_priv->dma_pages[i].flushed = 0; in savage_dma_reset()
362 if (dev_priv->dma_pages[page].age.wrap > wrap || in savage_dma_wait()
363 (dev_priv->dma_pages[pag in savage_dma_wait()
[all...]
H A Dsavage_drv.h178 drm_savage_dma_page_t *dma_pages; member
520 dev_priv->dma_pages[cur].used; \
526 dev_priv->dma_pages[cur].used; \
527 if (dev_priv->dma_pages[cur].used == 0) \
529 dev_priv->dma_pages[cur].used += (n); \
545 dev_priv->dma_pages[cur].used; \
/kernel/linux/linux-6.6/net/xdp/
H A Dxsk_buff_pool.c208 if (!pool->dma_pages) { in xp_assign_dev()
319 dma_map->dma_pages = kvcalloc(nr_pages, sizeof(*dma_map->dma_pages), GFP_KERNEL); in xp_create_dma_map()
320 if (!dma_map->dma_pages) { in xp_create_dma_map()
337 kvfree(dma_map->dma_pages); in xp_destroy_dma_map()
347 dma = &dma_map->dma_pages[i]; in __xp_dma_unmap()
363 if (!pool->dma_pages) in xp_dma_unmap()
376 kvfree(pool->dma_pages); in xp_dma_unmap()
377 pool->dma_pages = NULL; in xp_dma_unmap()
388 if (dma_map->dma_pages[ in xp_check_dma_contiguity()
[all...]
/kernel/linux/linux-5.10/net/xdp/
H A Dxsk_buff_pool.c175 if (!pool->dma_pages) { in __xp_assign_dev()
292 dma_map->dma_pages = kvcalloc(nr_pages, sizeof(*dma_map->dma_pages), GFP_KERNEL); in xp_create_dma_map()
293 if (!dma_map->dma_pages) { in xp_create_dma_map()
310 kvfree(dma_map->dma_pages); in xp_destroy_dma_map()
320 dma = &dma_map->dma_pages[i]; in __xp_dma_unmap()
349 kvfree(pool->dma_pages); in xp_dma_unmap()
360 if (dma_map->dma_pages[i] + PAGE_SIZE == dma_map->dma_pages[i + 1]) in xp_check_dma_contiguity()
361 dma_map->dma_pages[ in xp_check_dma_contiguity()
[all...]
/kernel/linux/linux-5.10/arch/c6x/mm/
H A Ddma-coherent.c33 static u32 dma_pages; variable
49 pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); in __alloc_dma_pages()
60 if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { in __free_dma_pages()
131 dma_pages = dma_size >> PAGE_SHIFT; in coherent_mem_init()
133 ++dma_pages; in coherent_mem_init()
135 dma_bitmap = memblock_alloc(BITS_TO_LONGS(dma_pages) * sizeof(long), in coherent_mem_init()
139 __func__, BITS_TO_LONGS(dma_pages) * sizeof(long), in coherent_mem_init()
/kernel/linux/linux-6.6/include/net/
H A Dxsk_buff_pool.h38 dma_addr_t *dma_pages; member
65 /* For performance reasons, each buff pool has its own array of dma_pages
68 dma_addr_t *dma_pages; member
126 dma_addr_t *dma_pages, u64 addr) in xp_init_xskb_dma()
128 xskb->frame_dma = (dma_pages[addr >> PAGE_SHIFT] & ~XSK_NEXT_PG_CONTIG_MASK) + in xp_init_xskb_dma()
185 return pool->dma_pages && in xp_desc_crosses_non_contig_pg()
186 !(pool->dma_pages[addr >> PAGE_SHIFT] & XSK_NEXT_PG_CONTIG_MASK); in xp_desc_crosses_non_contig_pg()
125 xp_init_xskb_dma(struct xdp_buff_xsk *xskb, struct xsk_buff_pool *pool, dma_addr_t *dma_pages, u64 addr) xp_init_xskb_dma() argument
/kernel/linux/linux-5.10/include/net/
H A Dxsk_buff_pool.h32 dma_addr_t *dma_pages; member
58 /* For performance reasons, each buff pool has its own array of dma_pages
61 dma_addr_t *dma_pages; member
154 !(pool->dma_pages[addr >> PAGE_SHIFT] & XSK_NEXT_PG_CONTIG_MASK); in xp_desc_crosses_non_contig_pg()

Completed in 6 milliseconds