/kernel/linux/linux-5.10/arch/alpha/kernel/ |
H A D | pci_iommu.c | 500 /* Classify the elements of the scatterlist. Write dma_address 537 sg->dma_address = -1; in sg_classify() 540 sg->dma_address = -2; in sg_classify() 545 leader->dma_address = leader_flag; in sg_classify() 555 leader->dma_address = leader_flag; in sg_classify() 576 if (leader->dma_address == 0 in sg_fill() 579 out->dma_address = paddr + __direct_map_base; in sg_fill() 583 __va(paddr), size, out->dma_address); in sg_fill() 590 if (leader->dma_address == 0 && dac_allowed) { in sg_fill() 591 out->dma_address in sg_fill() [all...] |
/kernel/linux/linux-6.6/arch/alpha/kernel/ |
H A D | pci_iommu.c | 473 /* Classify the elements of the scatterlist. Write dma_address 510 sg->dma_address = -1; in sg_classify() 513 sg->dma_address = -2; in sg_classify() 518 leader->dma_address = leader_flag; in sg_classify() 528 leader->dma_address = leader_flag; in sg_classify() 549 if (leader->dma_address == 0 in sg_fill() 552 out->dma_address = paddr + __direct_map_base; in sg_fill() 556 __va(paddr), size, out->dma_address); in sg_fill() 563 if (leader->dma_address == 0 && dac_allowed) { in sg_fill() 564 out->dma_address in sg_fill() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/ttm/ |
H A D | ttm_tt.c | 110 sizeof(*ttm->dma_address), GFP_KERNEL); in ttm_dma_tt_alloc_page_directory() 114 ttm->dma_address = (void *)(ttm->pages + ttm->num_pages); in ttm_dma_tt_alloc_page_directory() 120 ttm->dma_address = kvcalloc(ttm->num_pages, sizeof(*ttm->dma_address), in ttm_sg_tt_alloc_page_directory() 122 if (!ttm->dma_address) in ttm_sg_tt_alloc_page_directory() 141 ttm->dma_address = NULL; in ttm_tt_init_fields() 172 kvfree(ttm->dma_address); in ttm_tt_fini() 174 ttm->dma_address = NULL; in ttm_tt_fini()
|
/kernel/linux/linux-5.10/drivers/staging/gasket/ |
H A D | gasket_page_table.h | 218 * @dma_address: Dma address allocated by the kernel. 225 dma_addr_t *dma_address, uint64_t index); 228 dma_addr_t dma_address, uint64_t index); 238 * @dma_address: Dma address within the coherent memory range. 247 dma_addr_t dma_address, ulong vma);
|
H A D | gasket_ioctl.c | 196 ibuf.dma_address); in gasket_config_coherent_allocator() 206 ibuf.dma_address, in gasket_config_coherent_allocator() 210 &ibuf.dma_address, in gasket_config_coherent_allocator()
|
H A D | gasket.h | 43 * dma_address: phys addr start of coherent memory, allocated by kernel 49 u64 dma_address; member
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | amd_gart_64.c | 287 gart_unmap_page(dev, s->dma_address, s->dma_length, dir, 0); in gart_unmap_sg() 315 s->dma_address = addr; in dma_map_sg_nonforce() 338 unsigned long phys_addr = s->dma_address; in __dma_map_cont() 342 sout->dma_address = iommu_bus_base; in __dma_map_cont() 343 sout->dma_address += iommu_page*PAGE_SIZE + s->offset; in __dma_map_cont() 368 sout->dma_address = start->dma_address; in dma_map_cont() 402 s->dma_address = addr; in gart_map_sg() 460 s->dma_address = DMA_MAPPING_ERROR; in gart_map_sg()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/ttm/ |
H A D | ttm_tt.c | 100 sizeof(*ttm->dma_address), in ttm_dma_tt_alloc_page_directory() 104 ttm->dma_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages); in ttm_dma_tt_alloc_page_directory() 110 ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, in ttm_sg_tt_alloc_page_directory() 111 sizeof(*ttm->dma_address), in ttm_sg_tt_alloc_page_directory() 113 if (!ttm->dma_address) in ttm_sg_tt_alloc_page_directory() 303 kvfree(ttm_dma->dma_address); in ttm_dma_tt_fini() 305 ttm_dma->dma_address = NULL; in ttm_dma_tt_fini()
|
H A D | ttm_page_alloc.c | 1118 tt->dma_address[i] = dma_map_page(dev, tt->ttm.pages[i], in ttm_populate_and_map_pages() 1121 if (dma_mapping_error(dev, tt->dma_address[i])) { in ttm_populate_and_map_pages() 1123 dma_unmap_page(dev, tt->dma_address[i], in ttm_populate_and_map_pages() 1125 tt->dma_address[i] = 0; in ttm_populate_and_map_pages() 1132 tt->dma_address[i + 1] = tt->dma_address[i] + PAGE_SIZE; in ttm_populate_and_map_pages() 1148 if (!tt->dma_address[i] || !tt->ttm.pages[i]) { in ttm_unmap_and_unpopulate_pages() 1160 dma_unmap_page(dev, tt->dma_address[i], num_pages * PAGE_SIZE, in ttm_unmap_and_unpopulate_pages()
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | amd_gart_64.c | 285 gart_unmap_page(dev, s->dma_address, s->dma_length, dir, 0); in gart_unmap_sg() 313 s->dma_address = addr; in dma_map_sg_nonforce() 336 unsigned long phys_addr = s->dma_address; in __dma_map_cont() 340 sout->dma_address = iommu_bus_base; in __dma_map_cont() 341 sout->dma_address += iommu_page*PAGE_SIZE + s->offset; in __dma_map_cont() 366 sout->dma_address = start->dma_address; in dma_map_cont() 400 s->dma_address = addr; in gart_map_sg()
|
/kernel/linux/linux-5.10/arch/sparc/kernel/ |
H A D | iommu.c | 536 outs->dma_address = dma_addr; in dma_4u_map_sg() 549 outs->dma_address = DMA_MAPPING_ERROR; in dma_4u_map_sg() 561 vaddr = s->dma_address & IO_PAGE_MASK; in dma_4u_map_sg() 562 npages = iommu_num_pages(s->dma_address, s->dma_length, in dma_4u_map_sg() 575 s->dma_address = DMA_MAPPING_ERROR; in dma_4u_map_sg() 598 bus_addr = sg->dma_address & IO_PAGE_MASK; in fetch_sg_ctx() 627 dma_addr_t dma_handle = sg->dma_address; in dma_4u_unmap_sg() 722 iopte = iommu->page_table + ((sglist[0].dma_address - in dma_4u_sync_sg_for_cpu() 728 bus_addr = sglist[0].dma_address & IO_PAGE_MASK; in dma_4u_sync_sg_for_cpu() 736 npages = (IO_PAGE_ALIGN(sgprv->dma_address in dma_4u_sync_sg_for_cpu() [all...] |
/kernel/linux/linux-5.10/arch/s390/pci/ |
H A D | pci_dma.c | 495 s->dma_address = DMA_MAPPING_ERROR; in s390_dma_map_sg() 501 &dma->dma_address, dir)) in s390_dma_map_sg() 504 dma->dma_address += offset; in s390_dma_map_sg() 514 if (__s390_dma_map_sg(dev, start, size, &dma->dma_address, dir)) in s390_dma_map_sg() 517 dma->dma_address += offset; in s390_dma_map_sg() 538 s390_dma_unmap_pages(dev, s->dma_address, s->dma_length, in s390_dma_unmap_sg() 540 s->dma_address = 0; in s390_dma_unmap_sg()
|
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | swiotlb-xen.c | 499 xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg), in xen_swiotlb_unmap_sg() 514 sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg), in xen_swiotlb_map_sg() 516 if (sg->dma_address == DMA_MAPPING_ERROR) in xen_swiotlb_map_sg() 536 xen_swiotlb_sync_single_for_cpu(dev, sg->dma_address, in xen_swiotlb_sync_sg_for_cpu() 549 xen_swiotlb_sync_single_for_device(dev, sg->dma_address, in xen_swiotlb_sync_sg_for_device()
|
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | swiotlb-xen.c | 320 xen_swiotlb_unmap_page(hwdev, sg->dma_address, sg_dma_len(sg), in xen_swiotlb_unmap_sg() 335 sg->dma_address = xen_swiotlb_map_page(dev, sg_page(sg), in xen_swiotlb_map_sg() 337 if (sg->dma_address == DMA_MAPPING_ERROR) in xen_swiotlb_map_sg() 357 xen_swiotlb_sync_single_for_cpu(dev, sg->dma_address, in xen_swiotlb_sync_sg_for_cpu() 370 xen_swiotlb_sync_single_for_device(dev, sg->dma_address, in xen_swiotlb_sync_sg_for_device()
|
H A D | grant-dma-ops.c | 245 xen_grant_dma_unmap_page(dev, s->dma_address, sg_dma_len(s), dir, in xen_grant_dma_unmap_sg() 260 s->dma_address = xen_grant_dma_map_page(dev, sg_page(s), s->offset, in xen_grant_dma_map_sg() 262 if (s->dma_address == DMA_MAPPING_ERROR) in xen_grant_dma_map_sg()
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | pci-dma-compat.h | 58 pci_unmap_page(struct pci_dev *hwdev, dma_addr_t dma_address, in pci_unmap_page() argument 61 dma_unmap_page(&hwdev->dev, dma_address, size, (enum dma_data_direction)direction); in pci_unmap_page()
|
/kernel/linux/linux-6.6/arch/sparc/kernel/ |
H A D | iommu.c | 536 outs->dma_address = dma_addr; in dma_4u_map_sg() 560 vaddr = s->dma_address & IO_PAGE_MASK; in dma_4u_map_sg() 561 npages = iommu_num_pages(s->dma_address, s->dma_length, in dma_4u_map_sg() 596 bus_addr = sg->dma_address & IO_PAGE_MASK; in fetch_sg_ctx() 625 dma_addr_t dma_handle = sg->dma_address; in dma_4u_unmap_sg() 720 iopte = iommu->page_table + ((sglist[0].dma_address - in dma_4u_sync_sg_for_cpu() 726 bus_addr = sglist[0].dma_address & IO_PAGE_MASK; in dma_4u_sync_sg_for_cpu() 734 npages = (IO_PAGE_ALIGN(sgprv->dma_address + sgprv->dma_length) in dma_4u_sync_sg_for_cpu()
|
/kernel/linux/linux-6.6/arch/s390/pci/ |
H A D | pci_dma.c | 522 &dma->dma_address, dir); in s390_dma_map_sg() 526 dma->dma_address += offset; in s390_dma_map_sg() 536 ret = __s390_dma_map_sg(dev, start, size, &dma->dma_address, dir); in s390_dma_map_sg() 540 dma->dma_address += offset; in s390_dma_map_sg() 561 s390_dma_unmap_pages(dev, s->dma_address, s->dma_length, in s390_dma_unmap_sg() 563 s->dma_address = 0; in s390_dma_unmap_sg()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 296 dma_addr_t *dma_address; in amdgpu_ttm_map_buffer() local 299 dma_address = &dma->dma_address[offset >> PAGE_SHIFT]; in amdgpu_ttm_map_buffer() 300 r = amdgpu_gart_map(adev, 0, num_pages, dma_address, flags, in amdgpu_ttm_map_buffer() 305 dma_addr_t dma_address; in amdgpu_ttm_map_buffer() local 307 dma_address = (mm_node->start << PAGE_SHIFT) + offset; in amdgpu_ttm_map_buffer() 308 dma_address += adev->vm_manager.vram_base_offset; in amdgpu_ttm_map_buffer() 312 &dma_address, flags, cpu_addr); in amdgpu_ttm_map_buffer() 316 dma_address += PAGE_SIZE; in amdgpu_ttm_map_buffer() 1013 gtt->ttm.dma_address, tt in amdgpu_ttm_tt_pin_userptr() [all...] |
/kernel/linux/linux-5.10/arch/sparc/mm/ |
H A D | io-unit.c | 173 sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length); in iounit_map_sg() 205 len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_unmap_sg() 206 vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT; in iounit_unmap_sg()
|
H A D | iommu.c | 256 sg->dma_address =__sbus_iommu_map_page(dev, sg_page(sg), in __sbus_iommu_map_sg() 258 if (sg->dma_address == DMA_MAPPING_ERROR) in __sbus_iommu_map_sg() 305 sbus_iommu_unmap_page(dev, sg->dma_address, sg->length, dir, in sbus_iommu_unmap_sg() 307 sg->dma_address = 0x21212121; in sbus_iommu_unmap_sg()
|
/kernel/linux/linux-6.6/arch/sparc/mm/ |
H A D | io-unit.c | 174 sg->dma_address = iounit_get_area(iounit, (unsigned long) sg_virt(sg), sg->length); in iounit_map_sg() 206 len = ((sg->dma_address & ~PAGE_MASK) + sg->length + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_unmap_sg() 207 vaddr = (sg->dma_address - IOUNIT_DMA_BASE) >> PAGE_SHIFT; in iounit_unmap_sg()
|
H A D | iommu.c | 257 sg->dma_address =__sbus_iommu_map_page(dev, sg_page(sg), in __sbus_iommu_map_sg() 259 if (sg->dma_address == DMA_MAPPING_ERROR) in __sbus_iommu_map_sg() 306 sbus_iommu_unmap_page(dev, sg->dma_address, sg->length, dir, in sbus_iommu_unmap_sg() 308 sg->dma_address = 0x21212121; in sbus_iommu_unmap_sg()
|
/kernel/linux/linux-5.10/arch/ia64/hp/common/ |
H A D | sba_iommu.c | 391 startsg->dma_address, startsg->dma_length, in sba_dump_sg() 1214 nents, startsg->dma_address, cnt, in sba_fill_pdir() 1218 nents, startsg->dma_address, cnt, in sba_fill_pdir() 1224 if (startsg->dma_address & PIDE_FLAG) { in sba_fill_pdir() 1225 u32 pide = startsg->dma_address & ~PIDE_FLAG; in sba_fill_pdir() 1227 startsg->dma_address = 0; in sba_fill_pdir() 1230 dma_sg->dma_address = pide | ioc->ibase; in sba_fill_pdir() 1318 startsg->dma_address = startsg->dma_length = 0; in sba_coalesce_chunks() 1330 startsg->dma_address = startsg->dma_length = 0; in sba_coalesce_chunks() 1408 dma_sg->dma_address in sba_coalesce_chunks() [all...] |
/kernel/linux/linux-6.6/arch/ia64/hp/common/ |
H A D | sba_iommu.c | 391 startsg->dma_address, startsg->dma_length, in sba_dump_sg() 1222 nents, startsg->dma_address, cnt, in sba_fill_pdir() 1226 nents, startsg->dma_address, cnt, in sba_fill_pdir() 1232 if (startsg->dma_address & PIDE_FLAG) { in sba_fill_pdir() 1233 u32 pide = startsg->dma_address & ~PIDE_FLAG; in sba_fill_pdir() 1235 startsg->dma_address = 0; in sba_fill_pdir() 1238 dma_sg->dma_address = pide | ioc->ibase; in sba_fill_pdir() 1326 startsg->dma_address = startsg->dma_length = 0; in sba_coalesce_chunks() 1338 startsg->dma_address = startsg->dma_length = 0; in sba_coalesce_chunks() 1416 dma_sg->dma_address in sba_coalesce_chunks() [all...] |