/device/soc/hisilicon/common/platform/wifi/hi3881v100/driver/oal/ |
H A D | oal_scatterlist.h | 30 struct scatterlist *sgl; /* the list */ member 76 struct scatterlist *sgl = NULL; in __sg_free_table() local 79 if (unlikely(table->sgl == NULL)) { in __sg_free_table() 83 sgl = table->sgl; in __sg_free_table() 94 if (unlikely(sgl == NULL)) { in __sg_free_table() 98 next = sg_chain_ptr(&sgl[max_ents - 1]); in __sg_free_table() 107 free_fn(sgl, alloc_size); in __sg_free_table() 108 sgl = next; in __sg_free_table() 111 table->sgl in __sg_free_table() 125 sg_chain(struct scatterlist *prv, unsigned int prv_nents, struct scatterlist *sgl) sg_chain() argument [all...] |
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/ump/linux/ |
H A D | ump_ukk_ref_wrappers.c | 119 struct scatterlist *sgl; in ump_dmabuf_import_wrapper() local 168 for_each_sg(sgt->sgl, sgl, sgt->nents, i) { in ump_dmabuf_import_wrapper() 169 blocks[i].addr = sg_phys(sgl); in ump_dmabuf_import_wrapper() 170 blocks[i].size = sg_dma_len(sgl); in ump_dmabuf_import_wrapper()
|
/device/soc/rockchip/common/vendor/drivers/media/platform/rockchip/cif/ |
H A D | common.c | 36 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkcif_alloc_buffer() 115 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkcif_alloc_page_dummy_buf() 116 buf->dma_addr = sg_dma_address(sg->sgl); in rkcif_alloc_page_dummy_buf() 139 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkcif_free_page_dummy_buf()
|
/device/soc/rockchip/rk3588/kernel/drivers/media/platform/rockchip/cif/ |
H A D | common.c | 37 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkcif_alloc_buffer() 113 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkcif_alloc_page_dummy_buf() 114 buf->dma_addr = sg_dma_address(sg->sgl); in rkcif_alloc_page_dummy_buf() 137 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkcif_free_page_dummy_buf()
|
/device/soc/rockchip/rk3588/kernel/drivers/video/rockchip/rga3/ |
H A D | rga2_mmu_info.c | 363 struct scatterlist *sgl = sg->sgl; in rga2_MapION() local 368 len = sg_dma_len(sgl) >> PAGE_SHIFT; in rga2_MapION() 375 Address = sg_dma_address(sgl); in rga2_MapION() 389 } while ((sgl = sg_next(sgl)) && (mapped_size < pageCount) in rga2_MapION() 486 struct scatterlist *sgl = sg->sgl; in rga2_sgt_to_page_table() local 491 len = sg_dma_len(sgl) >> PAGE_SHIFT; in rga2_sgt_to_page_table() 493 len = sgl in rga2_sgt_to_page_table() [all...] |
H A D | rga_dma_buf.c | 472 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) in rga_iommu_map_virt_addr() 476 map_size = iommu_map_sg_atomic(domain, iova, sgt->sgl, sgt->orig_nents, ioprot); in rga_iommu_map_virt_addr() 596 for_each_sg(sgt.sgl, sg, sgt.orig_nents, i) in rga_viraddr_get_channel_info() 600 map_size = iommu_map_sg_atomic(domain, iova, sgt.sgl, sgt.orig_nents, ioprot); in rga_viraddr_get_channel_info() 827 rga_dma_buffer->iova = sg_dma_address(sgt->sgl); in rga_dma_map_fd() 828 rga_dma_buffer->size = sg_dma_len(sgt->sgl); in rga_dma_map_fd() 882 rga_dma_buffer->iova = sg_dma_address(sgt->sgl); in rga_dma_map_buffer() 885 rga_dma_buffer->size = sg_dma_len(sgt->sgl); in rga_dma_map_buffer()
|
H A D | rga_mm.c | 401 internal_buffer->dma_buffer[i].sgt->sgl, in rga_mm_unmap_virt_addr() 488 internal_buffer->dma_buffer[i].sgt->sgl, in rga_mm_map_virt_addr() 512 internal_buffer->dma_buffer[i].sgt->sgl, in rga_mm_map_virt_addr() 829 dma_sync_sg_for_device(scheduler->dev, sgt->sgl, sgt->orig_nents, dir); in rga_mm_sync_dma_sg_for_device() 855 dma_sync_sg_for_cpu(scheduler->dev, sgt->sgl, sgt->orig_nents, dir); in rga_mm_sync_dma_sg_for_cpu()
|
/device/soc/rockchip/common/sdk_linux/drivers/gpu/drm/rockchip/ |
H A D | rockchip_drm_gem.c | 286 for_each_sg(sgt->sgl, s, sgt->nents, i) sg_dma_address(s) = sg_phys(s); in rockchip_gem_alloc_dma() 819 rk_obj->dma_addr = sg_dma_address(sg->sgl); in rockchip_gem_dma_map_sg() 948 dma_sync_sg_for_cpu(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, dir); in rockchip_gem_prime_begin_cpu_access() 961 dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, dir); in rockchip_gem_prime_end_cpu_access() 965 static int rockchip_gem_prime_sgl_sync_range(struct device *dev, struct scatterlist *sgl, unsigned int nents, in rockchip_gem_prime_sgl_sync_range() argument 974 for_each_sg(sgl, sg, nents, i) in rockchip_gem_prime_sgl_sync_range() 1015 rockchip_gem_prime_sgl_sync_range(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, offset, len, dir, true); in rockchip_gem_prime_begin_cpu_access_partial() 1030 rockchip_gem_prime_sgl_sync_range(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, offset, len, dir, false); in rockchip_gem_prime_end_cpu_access_partial()
|
/device/soc/rockchip/common/sdk_linux/include/linux/ |
H A D | dmaengine.h | 140 * @src_sgl: If the 'icg' of sgl[] applies to Source (scattered read). 143 * @dst_sgl: If the 'icg' of sgl[] applies to Destination (scattered write). 147 * @frame_size: Number of chunks in a frame i.e, size of sgl[]. 148 * @sgl: Array of {chunk,icg} pairs that make up a frame. 160 struct data_chunk sgl[]; member 898 struct dma_async_tx_descriptor *(*device_prep_slave_sg)(struct dma_chan *chan, struct scatterlist *sgl, 958 static inline struct dma_async_tx_descriptor *dmaengine_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, in dmaengine_prep_slave_sg() argument 967 return chan->device->device_prep_slave_sg(chan, sgl, sg_len, dir, flags, NULL); in dmaengine_prep_slave_sg() 972 static inline struct dma_async_tx_descriptor *dmaengine_prep_rio_sg(struct dma_chan *chan, struct scatterlist *sgl, in dmaengine_prep_rio_sg() argument 981 return chan->device->device_prep_slave_sg(chan, sgl, sg_le in dmaengine_prep_rio_sg() 1271 dmaengine_get_icg(bool inc, bool sgl, size_t icg, size_t dir_icg) dmaengine_get_icg() argument [all...] |
H A D | dma-mapping.h | 317 nents = dma_map_sg_attrs(dev, sgt->sgl, sgt->orig_nents, dir, attrs); in dma_map_sgtable() 339 dma_unmap_sg_attrs(dev, sgt->sgl, sgt->orig_nents, dir, attrs); in dma_unmap_sgtable() 356 dma_sync_sg_for_cpu(dev, sgt->sgl, sgt->orig_nents, dir); in dma_sync_sgtable_for_cpu() 372 dma_sync_sg_for_device(dev, sgt->sgl, sgt->orig_nents, dir); in dma_sync_sgtable_for_device()
|
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory_secure.c | 78 for_each_sg(secure_mem->sgt->sgl, sg, secure_mem->sgt->nents, i) in mali_mem_secure_mali_map() 119 for_each_sg(secure_mem->sgt->sgl, sg, secure_mem->sgt->nents, i) in mali_mem_secure_cpu_map()
|
H A D | mali_memory_dma_buf.c | 91 for_each_sg(mem->sgt->sgl, sg, mem->sgt->nents, i) in mali_dma_buf_map() 109 guard_phys = sg_dma_address(mem->sgt->sgl); in mali_dma_buf_map()
|
/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
H A D | mali_memory_secure.c | 76 for_each_sg(secure_mem->sgt->sgl, sg, secure_mem->sgt->nents, i) { in mali_mem_secure_mali_map() 118 for_each_sg(secure_mem->sgt->sgl, sg, secure_mem->sgt->nents, i) { in mali_mem_secure_cpu_map()
|
H A D | mali_memory_dma_buf.c | 90 for_each_sg(mem->sgt->sgl, sg, mem->sgt->nents, i) { in mali_dma_buf_map() 107 guard_phys = sg_dma_address(mem->sgt->sgl); in mali_dma_buf_map()
|
/device/soc/rockchip/common/vendor/drivers/media/platform/rockchip/isp/ |
H A D | videobuf2-rdma-sg.c | 340 rd = buf->dma_sgt->sgl;
in vb2_dma_sg_dmabuf_ops_attach() 341 wr = sgt->sgl;
in vb2_dma_sg_dmabuf_ops_attach() 427 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->nents, buf->dma_dir);
in vb2_dma_sg_dmabuf_ops_begin_cpu_access() 436 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->nents, buf->dma_dir);
in vb2_dma_sg_dmabuf_ops_end_cpu_access()
|
H A D | common.c | 146 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkisp_alloc_buffer() 282 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkisp_alloc_page_dummy_buf() 283 dummy_buf->dma_addr = sg_dma_address(sg->sgl); in rkisp_alloc_page_dummy_buf() 307 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkisp_free_page_dummy_buf()
|
/device/soc/rockchip/rk3588/kernel/drivers/media/platform/rockchip/isp/ |
H A D | videobuf2-rdma-sg.c | 330 rd = buf->dma_sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach() 331 wr = sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach() 419 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->nents, buf->dma_dir); in vb2_dma_sg_dmabuf_ops_begin_cpu_access() 430 dma_sync_sg_for_device(buf->dev, sgt->sgl, sgt->nents, buf->dma_dir); in vb2_dma_sg_dmabuf_ops_end_cpu_access()
|
H A D | common.c | 144 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkisp_alloc_buffer() 276 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkisp_alloc_page_dummy_buf() 277 dummy_buf->dma_addr = sg_dma_address(sg->sgl); in rkisp_alloc_page_dummy_buf() 301 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkisp_free_page_dummy_buf()
|
/device/soc/rockchip/common/vendor/drivers/media/platform/rockchip/ispp/ |
H A D | common.c | 94 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkispp_allow_buffer() 328 pool->dma[i] = sg_dma_address(sg_tbl->sgl); in rkispp_init_pool() 452 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkispp_alloc_page_dummy_buf() 453 dummy_buf->dma_addr = sg_dma_address(sg->sgl); in rkispp_alloc_page_dummy_buf() 476 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkispp_free_page_dummy_buf()
|
/device/soc/rockchip/rk3588/kernel/drivers/media/platform/rockchip/ispp/ |
H A D | common.c | 92 buf->dma_addr = sg_dma_address(sg_tbl->sgl); in rkispp_allow_buffer() 320 pool->dma[i] = sg_dma_address(sg_tbl->sgl); in rkispp_init_pool() 437 ret = dma_map_sg(hw->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkispp_alloc_page_dummy_buf() 438 dummy_buf->dma_addr = sg_dma_address(sg->sgl); in rkispp_alloc_page_dummy_buf() 462 dma_unmap_sg(dev->hw_dev->dev, sg->sgl, sg->nents, DMA_BIDIRECTIONAL); in rkispp_free_page_dummy_buf()
|
/device/soc/rockchip/common/sdk_linux/drivers/iommu/ |
H A D | dma-iommu.c | 704 for_each_sg(sgt.sgl, sg, sgt.orig_nents, i) arch_dma_prep_coherent(sg_page(sg), sg->length); in iommu_dma_alloc_remap() 707 if (iommu_map_sg_atomic(domain, iova, sgt.sgl, sgt.orig_nents, ioprot) < size) { in iommu_dma_alloc_remap() 771 static void iommu_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nelems, in iommu_dma_sync_sg_for_cpu() argument 781 for_each_sg(sgl, sg, nelems, i) arch_sync_dma_for_cpu(sg_phys(sg), sg->length, dir); in iommu_dma_sync_sg_for_cpu() 784 static void iommu_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nelems, in iommu_dma_sync_sg_for_device() argument 794 for_each_sg(sgl, sg, nelems, i) arch_sync_dma_for_device(sg_phys(sg), sg->length, dir); in iommu_dma_sync_sg_for_device() 1218 sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0); in iommu_dma_get_sgtable()
|
/device/soc/rockchip/common/sdk_linux/drivers/dma-buf/heaps/ |
H A D | system_heap.c | 82 new_sg = new_table->sgl; in dup_sg_table() 355 for_each_sg(table->sgl, sg, table->nents, i) in system_heap_buf_free() 468 sg = table->sgl; in system_heap_do_allocate()
|
/device/soc/rockchip/common/vendor/drivers/video/rockchip/rga2/ |
H A D | rga2_mmu_info.c | 295 rga_dma_buffer->size = sg_dma_len(sgt->sgl);
in rga2_map_dma_buffer() 951 struct scatterlist *sgl = sg->sgl;
in rga2_MapION() local 958 len = sg_dma_len(sgl) >> PAGE_SHIFT;
in rga2_MapION() 963 Address = sg_dma_address(sgl);
in rga2_MapION() 977 } while ((sgl = sg_next(sgl)) && (mapped_size < pageCount) && (sg_num < sg->nents));
in rga2_MapION()
|
/device/soc/hisilicon/hi3516dv300/sdk_linux/drv/mpp/component/hifb/src/init/linux/ |
H A D | hifb_init.c | 374 sg_set_page(fb_table->sgl, fb_page, buf_attach->dmabuf->size, 0); in hifb_memblock_map() 375 for_each_sg(fb_table->sgl, sg_list, fb_table->nents, i) { in hifb_memblock_map()
|
/device/soc/rockchip/common/vendor/drivers/video/rockchip/mpp/ |
H A D | mpp_iommu.c | 222 buffer->iova = sg_dma_address(sgt->sgl); in mpp_dma_import_fd() 223 buffer->size = sg_dma_len(sgt->sgl); in mpp_dma_import_fd()
|