Home
last modified time | relevance | path

Searched refs:imem (Results 1 - 25 of 55) sorted by relevance

123

/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dgk20a.c54 struct gk20a_instmem *imem; member
145 struct gk20a_instmem *imem = obj->base.imem; in gk20a_instobj_iommu_recycle_vaddr() local
151 imem->vaddr_use -= nvkm_memory_size(&obj->base.memory); in gk20a_instobj_iommu_recycle_vaddr()
152 nvkm_debug(&imem->base.subdev, "vaddr used: %x/%x\n", imem->vaddr_use, in gk20a_instobj_iommu_recycle_vaddr()
153 imem->vaddr_max); in gk20a_instobj_iommu_recycle_vaddr()
160 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument
162 while (imem->vaddr_use + size > imem in gk20a_instmem_vaddr_gc()
177 struct gk20a_instmem *imem = node->imem; gk20a_instobj_acquire_dma() local
189 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_acquire_iommu() local
232 struct gk20a_instmem *imem = node->imem; gk20a_instobj_release_dma() local
244 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_release_iommu() local
298 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_dma() local
315 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_iommu() local
383 gk20a_instobj_ctor_dma(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_dma() argument
421 gk20a_instobj_ctor_iommu(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_iommu() argument
517 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instobj_new() local
550 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instmem_dtor() local
575 struct gk20a_instmem *imem; gk20a_instmem_new() local
[all...]
H A Dbase.c74 nvkm_instobj_dtor(struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_dtor() argument
76 spin_lock(&imem->lock); in nvkm_instobj_dtor()
78 spin_unlock(&imem->lock); in nvkm_instobj_dtor()
83 struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_ctor()
87 spin_lock(&imem->lock); in nvkm_instobj_ctor()
88 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_ctor()
89 spin_unlock(&imem->lock); in nvkm_instobj_ctor()
93 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, in nvkm_instobj_new() argument
96 struct nvkm_subdev *subdev = &imem->subdev; in nvkm_instobj_new()
101 ret = imem in nvkm_instobj_new()
82 nvkm_instobj_ctor(const struct nvkm_memory_func *func, struct nvkm_instmem *imem, struct nvkm_instobj *iobj) nvkm_instobj_ctor() argument
133 nvkm_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nvkm_instmem_rd32() argument
139 nvkm_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nvkm_instmem_wr32() argument
145 nvkm_instmem_boot(struct nvkm_instmem *imem) nvkm_instmem_boot() argument
162 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_fini() local
190 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_init() local
211 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_oneinit() local
220 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_dtor() local
235 nvkm_instmem_ctor(const struct nvkm_instmem_func *func, struct nvkm_device *device, int index, struct nvkm_instmem *imem) nvkm_instmem_ctor() argument
[all...]
H A Dnv04.c41 struct nv04_instmem *imem; member
49 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32()
57 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32()
76 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire()
102 mutex_lock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor()
103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor()
104 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv04_instobj_dtor()
105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv04_instobj_dtor()
123 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local
131 nvkm_instobj_ctor(&nv04_instobj_func, &imem in nv04_instobj_new()
147 nv04_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nv04_instmem_rd32() argument
153 nv04_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nv04_instmem_wr32() argument
161 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_oneinit() local
201 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_dtor() local
224 struct nv04_instmem *imem; nv04_instmem_new() local
[all...]
H A Dnv40.c43 struct nv40_instmem *imem; member
51 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32()
58 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32()
77 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire()
102 mutex_lock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor()
103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor()
104 mutex_unlock(&iobj->imem->base.subdev.mutex); in nv40_instobj_dtor()
105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv40_instobj_dtor()
123 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local
131 nvkm_instobj_ctor(&nv40_instobj_func, &imem in nv40_instobj_new()
161 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_oneinit() local
217 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_dtor() local
242 struct nv40_instmem *imem; nv40_instmem_new() local
[all...]
H A Dnv50.c47 struct nv50_instmem *imem; member
59 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_wr32_slow() local
60 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_wr32_slow()
65 spin_lock_irqsave(&imem->base.lock, flags); in nv50_instobj_wr32_slow()
66 if (unlikely(imem->addr != base)) { in nv50_instobj_wr32_slow()
68 imem->addr = base; in nv50_instobj_wr32_slow()
71 spin_unlock_irqrestore(&imem->base.lock, flags); in nv50_instobj_wr32_slow()
78 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32_slow() local
122 struct nv50_instmem *imem = iobj->imem; nv50_instobj_kmap() local
194 struct nv50_instmem *imem = iobj->imem; nv50_instobj_release() local
219 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_acquire() local
268 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_boot() local
319 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_dtor() local
359 struct nv50_instmem *imem = nv50_instmem(base); nv50_instobj_new() local
397 struct nv50_instmem *imem; nv50_instmem_new() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dgk20a.c54 struct gk20a_instmem *imem; member
145 struct gk20a_instmem *imem = obj->base.imem; in gk20a_instobj_iommu_recycle_vaddr() local
151 imem->vaddr_use -= nvkm_memory_size(&obj->base.memory); in gk20a_instobj_iommu_recycle_vaddr()
152 nvkm_debug(&imem->base.subdev, "vaddr used: %x/%x\n", imem->vaddr_use, in gk20a_instobj_iommu_recycle_vaddr()
153 imem->vaddr_max); in gk20a_instobj_iommu_recycle_vaddr()
160 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument
162 while (imem->vaddr_use + size > imem in gk20a_instmem_vaddr_gc()
177 struct gk20a_instmem *imem = node->imem; gk20a_instobj_acquire_dma() local
189 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_acquire_iommu() local
232 struct gk20a_instmem *imem = node->imem; gk20a_instobj_release_dma() local
244 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_release_iommu() local
298 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_dma() local
315 struct gk20a_instmem *imem = node->base.imem; gk20a_instobj_dtor_iommu() local
383 gk20a_instobj_ctor_dma(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_dma() argument
421 gk20a_instobj_ctor_iommu(struct gk20a_instmem *imem, u32 npages, u32 align, struct gk20a_instobj **_node) gk20a_instobj_ctor_iommu() argument
518 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instobj_new() local
551 struct gk20a_instmem *imem = gk20a_instmem(base); gk20a_instmem_dtor() local
576 struct gk20a_instmem *imem; gk20a_instmem_new() local
[all...]
H A Dbase.c74 nvkm_instobj_dtor(struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_dtor() argument
76 spin_lock(&imem->lock); in nvkm_instobj_dtor()
78 spin_unlock(&imem->lock); in nvkm_instobj_dtor()
83 struct nvkm_instmem *imem, struct nvkm_instobj *iobj) in nvkm_instobj_ctor()
87 spin_lock(&imem->lock); in nvkm_instobj_ctor()
88 list_add_tail(&iobj->head, &imem->list); in nvkm_instobj_ctor()
89 spin_unlock(&imem->lock); in nvkm_instobj_ctor()
96 struct nvkm_instmem *imem = device->imem; in nvkm_instobj_wrap() local
98 if (!imem in nvkm_instobj_wrap()
82 nvkm_instobj_ctor(const struct nvkm_memory_func *func, struct nvkm_instmem *imem, struct nvkm_instobj *iobj) nvkm_instobj_ctor() argument
105 nvkm_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, struct nvkm_memory **pmemory) nvkm_instobj_new() argument
145 nvkm_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nvkm_instmem_rd32() argument
151 nvkm_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nvkm_instmem_wr32() argument
157 nvkm_instmem_boot(struct nvkm_instmem *imem) nvkm_instmem_boot() argument
174 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_fini() local
202 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_init() local
223 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_oneinit() local
232 struct nvkm_instmem *imem = nvkm_instmem(subdev); nvkm_instmem_dtor() local
249 nvkm_instmem_ctor(const struct nvkm_instmem_func *func, struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_instmem *imem) nvkm_instmem_ctor() argument
[all...]
H A Dnv04.c41 struct nv04_instmem *imem; member
49 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32()
57 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32()
76 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire()
102 mutex_lock(&iobj->imem->base.mutex); in nv04_instobj_dtor()
103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor()
104 mutex_unlock(&iobj->imem->base.mutex); in nv04_instobj_dtor()
105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv04_instobj_dtor()
123 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local
131 nvkm_instobj_ctor(&nv04_instobj_func, &imem in nv04_instobj_new()
146 nv04_instmem_rd32(struct nvkm_instmem *imem, u32 addr) nv04_instmem_rd32() argument
152 nv04_instmem_wr32(struct nvkm_instmem *imem, u32 addr, u32 data) nv04_instmem_wr32() argument
160 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_oneinit() local
200 struct nv04_instmem *imem = nv04_instmem(base); nv04_instmem_dtor() local
223 struct nv04_instmem *imem; nv04_instmem_new() local
[all...]
H A Dnv40.c43 struct nv40_instmem *imem; member
51 iowrite32_native(data, iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_wr32()
58 return ioread32_native(iobj->imem->iomem + iobj->node->offset + offset); in nv40_instobj_rd32()
77 return iobj->imem->iomem + iobj->node->offset; in nv40_instobj_acquire()
102 mutex_lock(&iobj->imem->base.mutex); in nv40_instobj_dtor()
103 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv40_instobj_dtor()
104 mutex_unlock(&iobj->imem->base.mutex); in nv40_instobj_dtor()
105 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv40_instobj_dtor()
123 struct nv40_instmem *imem = nv40_instmem(base); in nv40_instobj_new() local
131 nvkm_instobj_ctor(&nv40_instobj_func, &imem in nv40_instobj_new()
160 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_oneinit() local
216 struct nv40_instmem *imem = nv40_instmem(base); nv40_instmem_dtor() local
241 struct nv40_instmem *imem; nv40_instmem_new() local
[all...]
H A Dnv50.c47 struct nv50_instmem *imem; member
59 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_wr32_slow() local
60 struct nvkm_device *device = imem->base.subdev.device; in nv50_instobj_wr32_slow()
65 spin_lock_irqsave(&imem->base.lock, flags); in nv50_instobj_wr32_slow()
66 if (unlikely(imem->addr != base)) { in nv50_instobj_wr32_slow()
68 imem->addr = base; in nv50_instobj_wr32_slow()
71 spin_unlock_irqrestore(&imem->base.lock, flags); in nv50_instobj_wr32_slow()
78 struct nv50_instmem *imem = iobj->imem; in nv50_instobj_rd32_slow() local
122 struct nv50_instmem *imem = iobj->imem; nv50_instobj_kmap() local
194 struct nv50_instmem *imem = iobj->imem; nv50_instobj_release() local
219 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_acquire() local
263 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_boot() local
314 struct nvkm_instmem *imem = &iobj->imem->base; nv50_instobj_dtor() local
354 struct nv50_instmem *imem = nv50_instmem(base); nv50_instobj_wrap() local
371 nv50_instobj_new(struct nvkm_instmem *imem, u32 size, u32 align, bool zero, struct nvkm_memory **pmemory) nv50_instobj_new() argument
409 struct nv50_instmem *imem; nv50_instmem_new() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Ddmanv40.c65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_fini() local
79 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_fini()
80 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, 0x00000000); in nv40_fifo_dma_engine_fini()
81 nvkm_done(imem->ramfc); in nv40_fifo_dma_engine_fini()
95 struct nvkm_instmem *imem = device->imem; in nv40_fifo_dma_engine_init() local
110 nvkm_kmap(imem->ramfc); in nv40_fifo_dma_engine_init()
111 nvkm_wo32(imem->ramfc, chan->ramfc + ctx, inst); in nv40_fifo_dma_engine_init()
112 nvkm_done(imem in nv40_fifo_dma_engine_init()
147 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; nv40_fifo_dma_object_ctor() local
193 struct nvkm_instmem *imem = device->imem; nv40_fifo_dma_new() local
[all...]
H A Ddmanv04.c39 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_dtor() local
42 nvkm_ramht_remove(imem->ramht, cookie); in nv04_fifo_dma_object_dtor()
51 struct nvkm_instmem *imem = chan->fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_object_ctor() local
67 hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, in nv04_fifo_dma_object_ctor()
79 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini()
144 struct nvkm_instmem *imem = fifo->base.engine.subdev.device->imem; in nv04_fifo_dma_dtor() local
147 nvkm_kmap(imem in nv04_fifo_dma_dtor()
175 struct nvkm_instmem *imem = device->imem; nv04_fifo_dma_new() local
[all...]
H A Ddmanv17.c46 struct nvkm_instmem *imem = device->imem; in nv17_fifo_dma_new() local
77 nvkm_kmap(imem->ramfc); in nv17_fifo_dma_new()
78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv17_fifo_dma_new()
79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv17_fifo_dma_new()
80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv17_fifo_dma_new()
81 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv17_fifo_dma_new()
88 nvkm_done(imem->ramfc); in nv17_fifo_dma_new()
H A Ddmanv10.c46 struct nvkm_instmem *imem = device->imem; in nv10_fifo_dma_new() local
76 nvkm_kmap(imem->ramfc); in nv10_fifo_dma_new()
77 nvkm_wo32(imem->ramfc, chan->ramfc + 0x00, args->v0.offset); in nv10_fifo_dma_new()
78 nvkm_wo32(imem->ramfc, chan->ramfc + 0x04, args->v0.offset); in nv10_fifo_dma_new()
79 nvkm_wo32(imem->ramfc, chan->ramfc + 0x0c, chan->base.push->addr >> 4); in nv10_fifo_dma_new()
80 nvkm_wo32(imem->ramfc, chan->ramfc + 0x14, in nv10_fifo_dma_new()
87 nvkm_done(imem->ramfc); in nv10_fifo_dma_new()
H A Dnv17.c55 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local
56 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init()
57 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init()
58 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
H A Dnv40.c65 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local
66 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init()
67 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init()
68 struct nvkm_memory *ramfc = imem->ramfc; in nv40_fifo_init()
/kernel/linux/linux-5.10/drivers/remoteproc/
H A Dqcom_pil_info.c33 struct resource imem; in qcom_pil_info_init() local
45 ret = of_address_to_resource(np, 0, &imem); in qcom_pil_info_init()
50 base = ioremap(imem.start, resource_size(&imem)); in qcom_pil_info_init()
56 memset_io(base, 0, resource_size(&imem)); in qcom_pil_info_init()
59 _reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE; in qcom_pil_info_init()
/kernel/linux/linux-6.6/drivers/remoteproc/
H A Dqcom_pil_info.c33 struct resource imem; in qcom_pil_info_init() local
45 ret = of_address_to_resource(np, 0, &imem); in qcom_pil_info_init()
50 base = ioremap(imem.start, resource_size(&imem)); in qcom_pil_info_init()
56 memset_io(base, 0, resource_size(&imem)); in qcom_pil_info_init()
59 _reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE; in qcom_pil_info_init()
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
H A Dnv40.c41 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv40_chan_ramfc_write()
114 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv40_eobj_ramht_add() local
119 hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context); in nv40_eobj_ramht_add()
129 struct nvkm_memory *ramfc = device->imem->ramfc; in nv40_ectx_bind()
185 struct nvkm_instmem *imem = device->imem; in nv40_fifo_init() local
186 struct nvkm_ramht *ramht = imem->ramht; in nv40_fifo_init()
187 struct nvkm_memory *ramro = imem->ramro; in nv40_fifo_init()
188 struct nvkm_memory *ramfc = imem in nv40_fifo_init()
[all...]
H A Dnv17.c40 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv17_chan_ramfc_write()
96 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local
97 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init()
98 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init()
99 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
H A Dnv04.c45 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_chan_stop()
105 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_clear()
118 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_write()
184 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv04_eobj_ramht_del() local
187 nvkm_ramht_remove(imem->ramht, hash); in nv04_eobj_ramht_del()
195 struct nvkm_instmem *imem = fifo->engine.subdev.device->imem; in nv04_eobj_ramht_add() local
200 hash = nvkm_ramht_insert(imem->ramht, eobj, chan->id, 4, eobj->handle, context); in nv04_eobj_ramht_add()
469 struct nvkm_instmem *imem in nv04_fifo_init() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/
H A Dnv40.c33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local
37 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma()
38 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma()
39 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/
H A Dnv40.c33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local
37 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma()
38 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma()
39 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/device/
H A Dbase.c86 .imem = nv04_instmem_new,
107 .imem = nv04_instmem_new,
129 .imem = nv04_instmem_new,
149 .imem = nv04_instmem_new,
171 .imem = nv04_instmem_new,
193 .imem = nv04_instmem_new,
215 .imem = nv04_instmem_new,
237 .imem = nv04_instmem_new,
259 .imem = nv04_instmem_new,
281 .imem
[all...]
/kernel/linux/linux-6.6/drivers/net/wwan/iosm/
H A Diosm_ipc_mux.c13 channel_id = ipc_imem_channel_alloc(ipc_mux->imem, ipc_mux->instance_id, in ipc_mux_channel_create()
25 ipc_mux->channel = ipc_imem_channel_open(ipc_mux->imem, channel_id, in ipc_mux_channel_create()
187 ipc_imem_channel_close(ipc_mux->imem, ipc_mux->channel_id); in ipc_mux_channel_close()
223 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule()
227 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule()
236 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule()
240 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule()
279 struct iosm_imem *imem) in ipc_mux_init()
295 ipc_mux->pcie = imem->pcie; in ipc_mux_init()
296 ipc_mux->imem in ipc_mux_init()
278 ipc_mux_init(struct ipc_mux_config *mux_cfg, struct iosm_imem *imem) ipc_mux_init() argument
[all...]

Completed in 15 milliseconds

123