Lines Matching refs:addr

71 	dmab->addr = 0;
234 return dmab->addr + offset;
280 static void *do_alloc_pages(struct device *dev, size_t size, dma_addr_t *addr,
290 *addr = page_to_phys(virt_to_page(p));
293 if ((*addr + size - 1) & ~dev->coherent_dma_mask) {
322 return do_alloc_pages(dmab->dev.dev, size, &dmab->addr, false);
334 dmab->addr >> PAGE_SHIFT,
384 unsigned long addr;
389 addr = get_vmalloc_page_addr(dmab, start);
394 addr += PAGE_SIZE;
395 if (get_vmalloc_page_addr(dmab, start) != addr)
427 p = gen_pool_dma_alloc_align(pool, size, &dmab->addr, PAGE_SIZE);
452 dmab->addr >> PAGE_SHIFT,
469 return dma_alloc_coherent(dmab->dev.dev, size, &dmab->addr, DEFAULT_GFP);
474 dma_free_coherent(dmab->dev.dev, dmab->bytes, dmab->area, dmab->addr);
481 dmab->area, dmab->addr, dmab->bytes);
497 return do_alloc_pages(dmab->dev.dev, size, &dmab->addr, true);
514 return dma_alloc_wc(dmab->dev.dev, size, &dmab->addr, DEFAULT_GFP);
519 dma_free_wc(dmab->dev.dev, dmab->bytes, dmab->area, dmab->addr);
526 dmab->area, dmab->addr, dmab->bytes);
563 dmab->addr = snd_sgbuf_get_addr(dmab, 0);
638 unsigned long addr;
646 addr = sg_page_iter_dma_address(&iter);
651 addr += PAGE_SIZE;
653 sg_page_iter_dma_address(&iter) != addr)
760 dma_addr_t addr;
786 p = dma_alloc_coherent(dmab->dev.dev, chunk, &addr, DEFAULT_GFP);
788 p = do_alloc_pages(dmab->dev.dev, chunk, &addr, false);
804 *addrp++ |= addr;
805 addr += PAGE_SIZE;
818 dmab->addr = sgbuf->addrs[0] & PAGE_MASK;
873 p = dma_alloc_noncoherent(dmab->dev.dev, size, &dmab->addr,
876 dmab->dev.need_sync = dma_need_sync(dmab->dev.dev, dmab->addr);
883 dmab->addr, dmab->dev.dir);
900 dma_sync_single_for_cpu(dmab->dev.dev, dmab->addr,
904 dma_sync_single_for_device(dmab->dev.dev, dmab->addr,