Lines Matching refs:dev_addr
285 dma_addr_t dev_addr;
316 dev_addr = xen_phys_to_dma(hwdev, phys);
317 if (((dev_addr + size - 1 <= dma_mask)) &&
319 *dma_handle = dev_addr;
335 dma_addr_t dev_addr, unsigned long attrs)
347 phys = xen_dma_to_phys(hwdev, dev_addr);
357 if (!WARN_ON((dev_addr + size - 1 > dma_mask) ||
379 dma_addr_t dev_addr = xen_phys_to_dma(dev, phys);
387 if (dma_capable(dev, dev_addr, size, true) &&
389 !xen_arch_need_swiotlb(dev, phys, dev_addr) &&
396 trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force);
403 dev_addr = xen_phys_to_dma(dev, map);
408 if (unlikely(!dma_capable(dev, dev_addr, size, true))) {
416 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr))))
419 xen_dma_sync_for_device(dev, dev_addr, size, dir);
421 return dev_addr;
432 static void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr,
435 phys_addr_t paddr = xen_dma_to_phys(hwdev, dev_addr);
440 if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr))))
443 xen_dma_sync_for_cpu(hwdev, dev_addr, size, dir);
446 /* NOTE: We use dev_addr here, not paddr! */
447 if (is_xen_swiotlb_buffer(hwdev, dev_addr))