/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | gk20a.c | 54 struct gk20a_instmem *imem; member 145 struct gk20a_instmem *imem = obj->base.imem; in gk20a_instobj_iommu_recycle_vaddr() local 151 imem->vaddr_use -= nvkm_memory_size(&obj->base.memory); in gk20a_instobj_iommu_recycle_vaddr() 152 nvkm_debug(&imem->base.subdev, "vaddr used: %x/%x\n", imem->vaddr_use, in gk20a_instobj_iommu_recycle_vaddr() 153 imem->vaddr_max); in gk20a_instobj_iommu_recycle_vaddr() 160 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument 162 while (imem->vaddr_use + size > imem in gk20a_instmem_vaddr_gc() 177 struct gk20a_instmem *imem = node->imem; gk20a_instobj_acquire_dma() local 189 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_acquire_iommu() local 232 struct gk20a_instmem *imem = node->imem; gk20a_instobj_release_dma() local 244 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_release_iommu() local 298 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_dma() local 315 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_iommu() local 383 gk20a_instobj_ctor_dma(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_dma() argument 421 gk20a_instobj_ctor_iommu(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_iommu() argument 517 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instobj_new() local 550 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instmem_dtor() local 575 struct gk20a_instmem *imem; gk20a_instmem_new() local [all...] |
H A D | base.c | 74 nvkm_instobj_dtor(struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_dtor() argument 76 spin_lock(&imem->lock); in nvkm_instobj_dtor() 78 spin_unlock(&imem->lock); in nvkm_instobj_dtor() 83 struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_ctor() 87 spin_lock(&imem->lock); in nvkm_instobj_ctor() 88 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_ctor() 89 spin_unlock(&imem->lock); in nvkm_instobj_ctor() 93 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, in nvkm_instobj_new() argument 96 struct nvkm_subdev *subdev = &imem->subdev; in nvkm_instobj_new() 101 ret = imem in nvkm_instobj_new() 82 nvkm_instobj_ctor(const struct nvkm_memory_func *func, struct nvkm_instmem *imem, struct nvkm_instobj *iobj) nvkm_instobj_ctor() argument 133 nvkm_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nvkm_instmem_rd32() argument 139 nvkm_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nvkm_instmem_wr32() argument 145 nvkm_instmem_boot(struct nvkm_instmem *imem) nvkm_instmem_boot() argument 162 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_fini() local 190 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_init() local 211 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_oneinit() local 220 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_dtor() local 235 nvkm_instmem_ctor(const struct nvkm_instmem_func *func, struct nvkm_device *device, int index, struct nvkm_instmem *imem) nvkm_instmem_ctor() argument [all...] |
H A D | nv04.c | 41 struct nv04_instmem *imem; member 49 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32() 57 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32() 76 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire() 102 mutex_lock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor() 103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor() 104 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor() 105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv04_instobj_dtor() 123 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local 131 nvkm_instobj_ctor(&nv04_instobj_func, &imem in nv04_instobj_new() 147 nv04_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nv04_instmem_rd32() argument 153 nv04_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nv04_instmem_wr32() argument 161 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_oneinit() local 201 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_dtor() local 224 struct nv04_instmem *imem; nv04_instmem_new() local [all...] |
H A D | nv40.c | 43 struct nv40_instmem *imem; member 51 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32() 58 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32() 77 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire() 102 mutex_lock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor() 103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor() 104 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor() 105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv40_instobj_dtor() 123 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local 131 nvkm_instobj_ctor(&nv40_instobj_func, &imem in nv40_instobj_new() 161 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_oneinit() local 217 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_dtor() local 242 struct nv40_instmem *imem; nv40_instmem_new() local [all...] |
H A D | nv50.c | 47 struct nv50_instmem *imem; member 59 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_wr32_slow() local 60 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_wr32_slow() 65 spin_lock_irqsave(&imem->base.lock, flags); in nv50_instobj_wr32_slow() 66 if (unlikely(imem->addr != base)) { in nv50_instobj_wr32_slow() 68 imem->addr = base; in nv50_instobj_wr32_slow() 71 spin_unlock_irqrestore(&imem->base.lock, flags); in nv50_instobj_wr32_slow() 78 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32_slow() local 122 struct nv50_instmem *imem = iobj->imem; nv50_instobj_kmap() local 194 struct nv50_instmem *imem = iobj->imem; nv50_instobj_release() local 219 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_acquire() local 268 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_boot() local 319 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_dtor() local 359 struct nv50_instmem *imem = nv50_instmem(base); nv50_instobj_new() local 397 struct nv50_instmem *imem; nv50_instmem_new() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | gk20a.c | 54 struct gk20a_instmem *imem; member 145 struct gk20a_instmem *imem = obj->base.imem; in gk20a_instobj_iommu_recycle_vaddr() local 151 imem->vaddr_use -= nvkm_memory_size(&obj->base.memory); in gk20a_instobj_iommu_recycle_vaddr() 152 nvkm_debug(&imem->base.subdev, "vaddr used: %x/%x\n", imem->vaddr_use, in gk20a_instobj_iommu_recycle_vaddr() 153 imem->vaddr_max); in gk20a_instobj_iommu_recycle_vaddr() 160 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument 162 while (imem->vaddr_use + size > imem in gk20a_instmem_vaddr_gc() 177 struct gk20a_instmem *imem = node->imem; gk20a_instobj_acquire_dma() local 189 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_acquire_iommu() local 232 struct gk20a_instmem *imem = node->imem; gk20a_instobj_release_dma() local 244 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_release_iommu() local 298 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_dma() local 315 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_iommu() local 383 gk20a_instobj_ctor_dma(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_dma() argument 421 gk20a_instobj_ctor_iommu(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_iommu() argument 518 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instobj_new() local 551 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instmem_dtor() local 576 struct gk20a_instmem *imem; gk20a_instmem_new() local [all...] |
H A D | base.c | 74 nvkm_instobj_dtor(struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_dtor() argument 76 spin_lock(&imem->lock); in nvkm_instobj_dtor() 78 spin_unlock(&imem->lock); in nvkm_instobj_dtor() 83 struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_ctor() 87 spin_lock(&imem->lock); in nvkm_instobj_ctor() 88 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_ctor() 89 spin_unlock(&imem->lock); in nvkm_instobj_ctor() 96 struct nvkm_instmem *imem = device->imem; in nvkm_instobj_wrap() local 98 if (!imem in nvkm_instobj_wrap() 82 nvkm_instobj_ctor(const struct nvkm_memory_func *func, struct nvkm_instmem *imem, struct nvkm_instobj *iobj) nvkm_instobj_ctor() argument 105 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, struct nvkm_memory **pmemory) nvkm_instobj_new() argument 145 nvkm_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nvkm_instmem_rd32() argument 151 nvkm_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nvkm_instmem_wr32() argument 157 nvkm_instmem_boot(struct nvkm_instmem *imem) nvkm_instmem_boot() argument 174 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_fini() local 202 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_init() local 223 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_oneinit() local 232 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_dtor() local 249 nvkm_instmem_ctor(const struct nvkm_instmem_func *func, struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_instmem *imem) nvkm_instmem_ctor() argument [all...] |
H A D | nv04.c | 41 struct nv04_instmem *imem; member 49 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32() 57 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32() 76 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire() 102 mutex_lock(&iobj->imem->base.mutex); in nv04_instobj_dtor() 103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor() 104 mutex_unlock(&iobj->imem->base.mutex); in nv04_instobj_dtor() 105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv04_instobj_dtor() 123 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local 131 nvkm_instobj_ctor(&nv04_instobj_func, &imem in nv04_instobj_new() 146 nv04_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nv04_instmem_rd32() argument 152 nv04_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nv04_instmem_wr32() argument 160 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_oneinit() local 200 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_dtor() local 223 struct nv04_instmem *imem; nv04_instmem_new() local [all...] |
H A D | nv40.c | 43 struct nv40_instmem *imem; member 51 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32() 58 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32() 77 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire() 102 mutex_lock(&iobj->imem->base.mutex); in nv40_instobj_dtor() 103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor() 104 mutex_unlock(&iobj->imem->base.mutex); in nv40_instobj_dtor() 105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv40_instobj_dtor() 123 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local 131 nvkm_instobj_ctor(&nv40_instobj_func, &imem in nv40_instobj_new() 160 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_oneinit() local 216 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_dtor() local 241 struct nv40_instmem *imem; nv40_instmem_new() local [all...] |
H A D | nv50.c | 47 struct nv50_instmem *imem; member 59 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_wr32_slow() local 60 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_wr32_slow() 65 spin_lock_irqsave(&imem->base.lock, flags); in nv50_instobj_wr32_slow() 66 if (unlikely(imem->addr != base)) { in nv50_instobj_wr32_slow() 68 imem->addr = base; in nv50_instobj_wr32_slow() 71 spin_unlock_irqrestore(&imem->base.lock, flags); in nv50_instobj_wr32_slow() 78 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32_slow() local 122 struct nv50_instmem *imem = iobj->imem; nv50_instobj_kmap() local 194 struct nv50_instmem *imem = iobj->imem; nv50_instobj_release() local 219 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_acquire() local 263 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_boot() local 314 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_dtor() local 354 struct nv50_instmem *imem = nv50_instmem(base); nv50_instobj_wrap() local 371 nv50_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, struct nvkm_memory **pmemory) nv50_instobj_new() argument 409 struct nv50_instmem *imem; nv50_instmem_new() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | dmanv40.c | 65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_fini() local 79 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_fini() 80 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, 0x00000000); in nv40_fifo_dma_engine_fini() 81 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_fini() 95 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_init() local 110 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_init() 111 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, inst); in nv40_fifo_dma_engine_init() 112 nvkm_done(imem in nv40_fifo_dma_engine_init() 147 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; nv40_fifo_dma_object_ctor() local 193 struct nvkm_instmem *imem = device->imem; nv40_fifo_dma_new() local [all...] |
H A D | dmanv04.c | 39 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_dtor() local 42 nvkm_ramht_remove(imem->ramht, cookie); in nv04_fifo_dma_object_dtor() 51 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_ctor() local 67 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv04_fifo_dma_object_ctor() 79 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini() 144 struct nvkm_instmem *imem = fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_dtor() local 147 nvkm_kmap(imem in nv04_fifo_dma_dtor() 175 struct nvkm_instmem *imem = device->imem; nv04_fifo_dma_new() local [all...] |
H A D | dmanv17.c | 46 struct nvkm_instmem *imem = device->imem; in nv17_fifo_dma_new() local 77 nvkm_kmap(imem->ramfc); in nv17_fifo_dma_new() 78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv17_fifo_dma_new() 79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv17_fifo_dma_new() 80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv17_fifo_dma_new() 81 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv17_fifo_dma_new() 88 nvkm_done(imem->ramfc); in nv17_fifo_dma_new()
|
H A D | dmanv10.c | 46 struct nvkm_instmem *imem = device->imem; in nv10_fifo_dma_new() local 76 nvkm_kmap(imem->ramfc); in nv10_fifo_dma_new() 77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv10_fifo_dma_new() 78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv10_fifo_dma_new() 79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv10_fifo_dma_new() 80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv10_fifo_dma_new() 87 nvkm_done(imem->ramfc); in nv10_fifo_dma_new()
|
H A D | nv17.c | 55 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local 56 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init() 57 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init() 58 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
|
H A D | nv40.c | 65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local 66 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init() 67 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init() 68 struct nvkm_memory *ramfc = imem->ramfc; in nv40_fifo_init()
|
/kernel/linux/linux-5.10/drivers/remoteproc/ |
H A D | qcom_pil_info.c | 33 struct resource imem; in qcom_pil_info_init() local 45 ret = of_address_to_resource(np, 0, &imem); in qcom_pil_info_init() 50 base = ioremap(imem.start, resource_size(&imem)); in qcom_pil_info_init() 56 memset_io(base, 0, resource_size(&imem)); in qcom_pil_info_init() 59 _reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE; in qcom_pil_info_init()
|
/kernel/linux/linux-6.6/drivers/remoteproc/ |
H A D | qcom_pil_info.c | 33 struct resource imem; in qcom_pil_info_init() local 45 ret = of_address_to_resource(np, 0, &imem); in qcom_pil_info_init() 50 base = ioremap(imem.start, resource_size(&imem)); in qcom_pil_info_init() 56 memset_io(base, 0, resource_size(&imem)); in qcom_pil_info_init() 59 _reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE; in qcom_pil_info_init()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | nv40.c | 41 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv40_chan_ramfc_write() 114 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv40_eobj_ramht_add() local 119 hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context); in nv40_eobj_ramht_add() 129 struct nvkm_memory *ramfc = device->imem->ramfc; in nv40_ectx_bind() 185 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local 186 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init() 187 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init() 188 struct nvkm_memory *ramfc = imem in nv40_fifo_init() [all...] |
H A D | nv17.c | 40 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv17_chan_ramfc_write() 96 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local 97 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init() 98 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init() 99 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
|
H A D | nv04.c | 45 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_chan_stop() 105 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_clear() 118 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_write() 184 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv04_eobj_ramht_del() local 187 nvkm_ramht_remove(imem->ramht, hash); in nv04_eobj_ramht_del() 195 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv04_eobj_ramht_add() local 200 hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context); in nv04_eobj_ramht_add() 469 struct nvkm_instmem *imem in nv04_fifo_init() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/ |
H A D | nv40.c | 33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local 37 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma() 38 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma() 39 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/ |
H A D | nv40.c | 33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local 37 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma() 38 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma() 39 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/device/ |
H A D | base.c | 86 .imem = nv04_instmem_new, 107 .imem = nv04_instmem_new, 129 .imem = nv04_instmem_new, 149 .imem = nv04_instmem_new, 171 .imem = nv04_instmem_new, 193 .imem = nv04_instmem_new, 215 .imem = nv04_instmem_new, 237 .imem = nv04_instmem_new, 259 .imem = nv04_instmem_new, 281 .imem [all...] |
/kernel/linux/linux-6.6/drivers/net/wwan/iosm/ |
H A D | iosm_ipc_mux.c | 13 channel_id = ipc_imem_channel_alloc(ipc_mux->imem, ipc_mux->instance_id, in ipc_mux_channel_create() 25 ipc_mux->channel = ipc_imem_channel_open(ipc_mux->imem, channel_id, in ipc_mux_channel_create() 187 ipc_imem_channel_close(ipc_mux->imem, ipc_mux->channel_id); in ipc_mux_channel_close() 223 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule() 227 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule() 236 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule() 240 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule() 279 struct iosm_imem *imem) in ipc_mux_init() 295 ipc_mux->pcie = imem->pcie; in ipc_mux_init() 296 ipc_mux->imem in ipc_mux_init() 278 ipc_mux_init(struct ipc_mux_config *mux_cfg, struct iosm_imem *imem) ipc_mux_init() argument [all...] |