/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | runl.c | 22 #include "runl.h" 46 chan = nvkm_runl_chan_get_chid(engn->runl, id, pirqflags); in nvkm_engn_cgrp_get() 50 cgrp = nvkm_runl_cgrp_get_cgid(engn->runl, id, pirqflags); in nvkm_engn_cgrp_get() 58 nvkm_runl_rc(struct nvkm_runl *runl) in nvkm_runl_rc() argument 60 struct nvkm_fifo *fifo = runl->fifo; in nvkm_runl_rc() 69 BUG_ON(!mutex_is_locked(&runl->mutex)); in nvkm_runl_rc() 70 rc = atomic_xchg(&runl->rc_pending, 0); in nvkm_runl_rc() 75 nvkm_runl_foreach_cgrp_safe(cgrp, gtmp, runl) { in nvkm_runl_rc() 87 if (runl->func->preempt) { in nvkm_runl_rc() 88 for (i = 0; i < runl in nvkm_runl_rc() 132 nvkm_runl_rc_runl(struct nvkm_runl *runl) nvkm_runl_rc_runl() argument 155 nvkm_runl_rc_engn(struct nvkm_runl *runl, struct nvkm_engn *engn) nvkm_runl_rc_engn() argument 174 struct nvkm_runl *runl = container_of(work, typeof(*runl), work); nvkm_runl_work() local 183 nvkm_runl_chan_get_inst(struct nvkm_runl *runl, u64 inst, unsigned long *pirqflags) nvkm_runl_chan_get_inst() argument 207 nvkm_runl_chan_get_chid(struct nvkm_runl *runl, int id, unsigned long *pirqflags) nvkm_runl_chan_get_chid() argument 228 nvkm_runl_cgrp_get_cgid(struct nvkm_runl *runl, int id, unsigned long *pirqflags) nvkm_runl_cgrp_get_cgid() argument 249 nvkm_runl_preempt_wait(struct nvkm_runl *runl) nvkm_runl_preempt_wait() argument 261 nvkm_runl_update_pending(struct nvkm_runl *runl) nvkm_runl_update_pending() argument 271 nvkm_runl_update_locked(struct nvkm_runl *runl, bool wait) nvkm_runl_update_locked() argument 281 nvkm_runl_allow(struct nvkm_runl *runl) nvkm_runl_allow() argument 295 nvkm_runl_block(struct nvkm_runl *runl) nvkm_runl_block() argument 309 nvkm_runl_fini(struct nvkm_runl *runl) nvkm_runl_fini() argument 318 nvkm_runl_del(struct nvkm_runl *runl) nvkm_runl_del() argument 338 nvkm_runl_add(struct nvkm_runl *runl, int engi, const struct nvkm_engn_func *func, enum nvkm_subdev_type type, int inst) nvkm_runl_add() argument 384 struct nvkm_runl *runl; nvkm_runl_get() local 398 struct nvkm_runl *runl; nvkm_runl_new() local [all...] |
H A D | ga100.c | 26 #include "runl.h" 38 return (chan->cgrp->runl->doorbell << 16) | chan->id; in ga100_chan_doorbell_handle() 44 struct nvkm_runl *runl = chan->cgrp->runl; in ga100_chan_stop() local 46 nvkm_wr32(runl->fifo->engine.subdev.device, runl->chan + (chan->id * 4), 0x00000003); in ga100_chan_stop() 52 struct nvkm_runl *runl = chan->cgrp->runl; in ga100_chan_start() local 53 struct nvkm_device *device = runl->fifo->engine.subdev.device; in ga100_chan_start() 56 nvkm_wr32(device, runl in ga100_chan_start() 63 struct nvkm_runl *runl = chan->cgrp->runl; ga100_chan_unbind() local 111 struct nvkm_runl *runl = cgrp->runl; ga100_cgrp_preempt() local 124 struct nvkm_runl *runl = engn->runl; ga100_engn_cxid() local 184 ga100_runq_intr_1(struct nvkm_runq *runq, struct nvkm_runl *runl) ga100_runq_intr_1() argument 222 ga100_runq_intr_0(struct nvkm_runq *runq, struct nvkm_runl *runl) ga100_runq_intr_0() argument 260 ga100_runq_intr(struct nvkm_runq *runq, struct nvkm_runl *runl) ga100_runq_intr() argument 287 ga100_runl_preempt_pending(struct nvkm_runl *runl) ga100_runl_preempt_pending() argument 293 ga100_runl_preempt(struct nvkm_runl *runl) ga100_runl_preempt() argument 299 ga100_runl_allow(struct nvkm_runl *runl, u32 engm) ga100_runl_allow() argument 305 ga100_runl_block(struct nvkm_runl *runl, u32 engm) ga100_runl_block() argument 311 ga100_runl_pending(struct nvkm_runl *runl) ga100_runl_pending() argument 319 ga100_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count) ga100_runl_commit() argument 332 struct nvkm_runl *runl = container_of(inth, typeof(*runl), inth); ga100_runl_intr() local 385 ga100_runl_fini(struct nvkm_runl *runl) ga100_runl_fini() argument 392 ga100_runl_init(struct nvkm_runl *runl) ga100_runl_init() argument 443 struct nvkm_runl *runl; ga100_runl_new() local 510 struct nvkm_runl *runl = container_of(inth, typeof(*runl), nonstall.inth); ga100_fifo_nonstall_intr() local 520 struct nvkm_runl *runl = nvkm_runl_get(fifo, index, 0); ga100_fifo_nonstall_block() local 529 struct nvkm_runl *runl = nvkm_runl_get(fifo, index, 0); ga100_fifo_nonstall_allow() local 545 struct nvkm_runl *runl; ga100_fifo_nonstall_ctor() local 573 struct nvkm_runl *runl; ga100_fifo_runl_ctor() local [all...] |
H A D | chan.c | 28 #include "runl.h" 45 struct nvkm_runl *runl = cgrp->runl; in nvkm_chan_cctx_bind() local 57 nvkm_runl_block(runl); in nvkm_chan_cctx_bind() 67 nvkm_runl_allow(runl); in nvkm_chan_cctx_bind() 139 struct nvkm_runl *runl = chan->cgrp->runl; in nvkm_chan_preempt_locked() local 146 return nvkm_runl_preempt_wait(runl); in nvkm_chan_preempt_locked() 157 mutex_lock(&chan->cgrp->runl->mutex); in nvkm_chan_preempt() 159 mutex_unlock(&chan->cgrp->runl in nvkm_chan_preempt() 167 struct nvkm_runl *runl = cgrp->runl; nvkm_chan_remove_locked() local 185 struct nvkm_runl *runl = chan->cgrp->runl; nvkm_chan_remove() local 199 struct nvkm_runl *runl = cgrp->runl; nvkm_chan_insert() local 311 struct nvkm_runl *runl; nvkm_chan_get_inst() local 332 struct nvkm_runl *runl; nvkm_chan_get_chid() local 346 nvkm_chan_new_(const struct nvkm_chan_func *func, struct nvkm_runl *runl, int runq, struct nvkm_cgrp *cgrp, const char *name, bool priv, u32 devm, struct nvkm_vmm *vmm, struct nvkm_dmaobj *dmaobj, u64 offset, u64 length, struct nvkm_memory *userd, u64 ouserd, struct nvkm_chan **pchan) nvkm_chan_new_() argument [all...] |
H A D | nv50.c | 28 #include "runl.h" 50 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in nv50_chan_stop() 58 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in nv50_chan_start() 66 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in nv50_chan_unbind() 74 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in nv50_chan_bind() 82 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in nv50_chan_ramfc_write() 152 struct nvkm_subdev *subdev = &chan->cgrp->runl->fifo->engine.subdev; in nv50_ectx_bind() 218 nv50_runl_pending(struct nvkm_runl *runl) in nv50_runl_pending() argument 220 return nvkm_rd32(runl->fifo->engine.subdev.device, 0x0032ec) & 0x00000100; in nv50_runl_pending() 224 nv50_runl_wait(struct nvkm_runl *runl) in nv50_runl_wait() argument 238 nv50_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count) nv50_runl_commit() argument 254 nv50_runl_alloc(struct nvkm_runl *runl, u32 *offset) nv50_runl_alloc() argument 285 nv50_runl_update(struct nvkm_runl *runl) nv50_runl_update() argument 341 struct nvkm_runl *runl = nvkm_runl_first(fifo); nv50_fifo_init() local [all...] |
H A D | g98.c | 24 #include "runl.h" 31 struct nvkm_runl *runl; in g98_fifo_runl_ctor() local 33 runl = nvkm_runl_new(fifo, 0, 0, 0); in g98_fifo_runl_ctor() 34 if (IS_ERR(runl)) in g98_fifo_runl_ctor() 35 return PTR_ERR(runl); in g98_fifo_runl_ctor() 37 nvkm_runl_add(runl, 0, fifo->func->engn_sw, NVKM_ENGINE_SW, 0); in g98_fifo_runl_ctor() 38 nvkm_runl_add(runl, 0, fifo->func->engn_sw, NVKM_ENGINE_DMAOBJ, 0); in g98_fifo_runl_ctor() 39 nvkm_runl_add(runl, 1, fifo->func->engn, NVKM_ENGINE_GR, 0); in g98_fifo_runl_ctor() 40 nvkm_runl_add(runl, 2, fifo->func->engn, NVKM_ENGINE_MSPPP, 0); in g98_fifo_runl_ctor() 41 nvkm_runl_add(runl, in g98_fifo_runl_ctor() [all...] |
H A D | runl.h | 24 struct nvkm_runl *runl; member 35 RUNL_PRINT((e)->runl, l, p, "%02d[%8s]:"f, (e)->id, (e)->engine->subdev.name, ##a) 114 #define nvkm_runl_find_engn(engn,runl,cond) nvkm_list_find(engn, &(runl)->engns, head, (cond)) 117 #define nvkm_runl_foreach(runl,fifo) list_for_each_entry((runl), &(fifo)->runls, head) 118 #define nvkm_runl_foreach_cond(runl,fifo,cond) nvkm_list_foreach(runl, &(fifo)->runls, head, (cond)) 119 #define nvkm_runl_foreach_engn(engn,runl) list_for_each_entry((engn), &(runl) [all...] |
H A D | base.c | 27 #include "runl.h" 41 struct nvkm_runl *runl; in nvkm_fifo_ctxsw_in_progress() local 44 nvkm_runl_foreach(runl, engine->subdev.device->fifo) { in nvkm_fifo_ctxsw_in_progress() 45 nvkm_runl_foreach_engn(engn, runl) { in nvkm_fifo_ctxsw_in_progress() 128 struct nvkm_runl *runl; in nvkm_fifo_fini() local 132 nvkm_runl_foreach(runl, fifo) in nvkm_fifo_fini() 133 nvkm_runl_fini(runl); in nvkm_fifo_fini() 143 struct nvkm_runl *runl; in nvkm_fifo_init() local 156 nvkm_runl_foreach(runl, fifo) { in nvkm_fifo_init() 157 if (runl in nvkm_fifo_init() 172 struct nvkm_runl *runl; nvkm_fifo_info() local 244 struct nvkm_runl *runl; nvkm_fifo_oneinit() local 335 struct nvkm_runl *runl, *runt; nvkm_fifo_dtor() local [all...] |
H A D | gk104.c | 28 #include "runl.h" 42 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gk104_chan_stop() 50 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gk104_chan_start() 58 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gk104_chan_unbind() 66 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gk104_chan_bind_inst() 74 struct nvkm_runl *runl = chan->cgrp->runl; in gk104_chan_bind() local 75 struct nvkm_device *device = runl->fifo->engine.subdev.device; in gk104_chan_bind() 77 nvkm_mask(device, 0x800004 + (chan->id * 8), 0x000f0000, runl->id << 16); in gk104_chan_bind() 206 u32 stat = nvkm_rd32(engn->runl in gk104_engn_status() 404 gk104_runl_fault_clear(struct nvkm_runl *runl) gk104_runl_fault_clear() argument 410 gk104_runl_allow(struct nvkm_runl *runl, u32 engm) gk104_runl_allow() argument 416 gk104_runl_block(struct nvkm_runl *runl, u32 engm) gk104_runl_block() argument 422 gk104_runl_pending(struct nvkm_runl *runl) gk104_runl_pending() argument 430 gk104_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count) gk104_runl_commit() argument 643 struct nvkm_runl *runl; gk104_fifo_intr_runlist() local 760 struct nvkm_runl *runl; gk104_fifo_runl_ctor() local [all...] |
H A D | tu102.c | 25 #include "runl.h" 36 return (chan->cgrp->runl->id << 16) | chan->id; in tu102_chan_doorbell_handle() 42 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in tu102_chan_start() 62 tu102_runl_pending(struct nvkm_runl *runl) in tu102_runl_pending() argument 64 struct nvkm_device *device = runl->fifo->engine.subdev.device; in tu102_runl_pending() 66 return nvkm_rd32(device, 0x002b0c + (runl->id * 0x10)) & 0x00008000; in tu102_runl_pending() 70 tu102_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count) in tu102_runl_commit() argument 72 struct nvkm_device *device = runl->fifo->engine.subdev.device; in tu102_runl_commit() 76 nvkm_wr32(device, 0x002b00 + (runl->id * 0x10), lower_32_bits(addr)); in tu102_runl_commit() 77 nvkm_wr32(device, 0x002b04 + (runl in tu102_runl_commit() 138 struct nvkm_runl *runl = engn->runl; tu102_fifo_intr_ctxsw_timeout_info() local 171 struct nvkm_runl *runl; tu102_fifo_intr_ctxsw_timeout() local [all...] |
H A D | g84.c | 27 #include "runl.h" 37 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in g84_chan_bind() 45 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in g84_chan_ramfc_write() 108 struct nvkm_subdev *subdev = &chan->cgrp->runl->fifo->engine.subdev; in g84_ectx_bind() 192 struct nvkm_runl *runl; in g84_fifo_runl_ctor() local 194 runl = nvkm_runl_new(fifo, 0, 0, 0); in g84_fifo_runl_ctor() 195 if (IS_ERR(runl)) in g84_fifo_runl_ctor() 196 return PTR_ERR(runl); in g84_fifo_runl_ctor() 198 nvkm_runl_add(runl, 0, fifo->func->engn_sw, NVKM_ENGINE_SW, 0); in g84_fifo_runl_ctor() 199 nvkm_runl_add(runl, in g84_fifo_runl_ctor() [all...] |
H A D | gf100.c | 28 #include "runl.h" 43 nvkm_wr32(chan->cgrp->runl->fifo->engine.subdev.device, 0x002634, chan->id); in gf100_chan_preempt() 49 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gf100_chan_stop() 57 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gf100_chan_start() 67 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in gf100_chan_unbind() 79 struct nvkm_device *device = chan->cgrp->runl->fifo->engine.subdev.device; in gf100_chan_bind() 204 struct nvkm_runl *runl = engn->runl; in gf100_engn_mmu_fault_triggered() local 205 struct nvkm_fifo *fifo = runl->fifo; in gf100_engn_mmu_fault_triggered() 215 if (atomic_dec_and_test(&runl in gf100_engn_mmu_fault_triggered() 224 struct nvkm_runl *runl = engn->runl; gf100_engn_mmu_fault_trigger() local 370 gf100_runl_preempt_pending(struct nvkm_runl *runl) gf100_runl_preempt_pending() argument 376 gf100_runl_fault_clear(struct nvkm_runl *runl) gf100_runl_fault_clear() argument 382 gf100_runl_allow(struct nvkm_runl *runl, u32 engm) gf100_runl_allow() argument 388 gf100_runl_block(struct nvkm_runl *runl, u32 engm) gf100_runl_block() argument 394 gf100_runl_pending(struct nvkm_runl *runl) gf100_runl_pending() argument 400 gf100_runl_commit(struct nvkm_runl *runl, struct nvkm_memory *memory, u32 start, int count) gf100_runl_commit() argument 540 struct nvkm_runl *runl; gf100_fifo_mmu_fault_recover() local 620 struct nvkm_runl *runl; gf100_fifo_intr_ctxsw_timeout() local 647 struct nvkm_runl *runl; gf100_fifo_intr_sched_ctxsw() local 910 struct nvkm_runl *runl; gf100_fifo_runl_ctor() local [all...] |
H A D | cgrp.c | 25 #include "runl.h" 175 struct nvkm_runl *runl = cgrp->runl; in nvkm_cgrp_del() local 177 if (runl->cgid) in nvkm_cgrp_del() 178 nvkm_chid_put(runl->cgid, cgrp->id, &cgrp->lock); in nvkm_cgrp_del() 219 nvkm_cgrp_new(struct nvkm_runl *runl, const char *name, struct nvkm_vmm *vmm, bool hw, in nvkm_cgrp_new() argument 227 cgrp->func = runl->fifo->func->cgrp.func; in nvkm_cgrp_new() 229 cgrp->runl = runl; in nvkm_cgrp_new() 242 if (runl in nvkm_cgrp_new() [all...] |
H A D | ucgrp.c | 25 #include "runl.h" 42 return nvkm_uchan_new(cgrp->runl->fifo, cgrp, oclass, argv, argc, pobject); in nvkm_ucgrp_chan_new() 49 struct nvkm_fifo *fifo = cgrp->runl->fifo; in nvkm_ucgrp_sclass() 85 struct nvkm_runl *runl; in nvkm_ucgrp_new() local 98 runl = nvkm_runl_get(fifo, args->v0.runlist, 0); in nvkm_ucgrp_new() 99 if (!runl) in nvkm_ucgrp_new() 115 ret = nvkm_cgrp_new(runl, args->v0.name, vmm, true, &ucgrp->cgrp); in nvkm_ucgrp_new()
|
H A D | uchan.c | 27 #include "runl.h" 45 struct nvkm_runl *runl = chan->cgrp->runl; in nvkm_uchan_uevent() local 55 return nvkm_uevent_add(uevent, &runl->fifo->nonstall.event, runl->id, in nvkm_uchan_uevent() 58 return nvkm_uevent_add(uevent, &runl->chid->event, chan->id, in nvkm_uchan_uevent() 164 engn = nvkm_runl_find_engn(engn, cgrp->runl, engn->engine == oclass->engine); in nvkm_uchan_object_new() 210 nvkm_runl_foreach_engn(engn, chan->cgrp->runl) { in nvkm_uchan_sclass() 215 if (engn->runl->func->runqs) { in nvkm_uchan_sclass() 259 struct nvkm_device *device = chan->cgrp->runl in nvkm_uchan_map() 325 struct nvkm_runl *runl; nvkm_uchan_new() local [all...] |
H A D | nv04.c | 28 #include "runl.h" 43 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv04_chan_stop() 94 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv04_chan_start() 105 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_clear() 118 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv04_chan_ramfc_write() 183 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv04_eobj_ramht_del() 194 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv04_eobj_ramht_add() 496 struct nvkm_runl *runl; in nv04_fifo_runl_ctor() local 498 runl = nvkm_runl_new(fifo, 0, 0, 0); in nv04_fifo_runl_ctor() 499 if (IS_ERR(runl)) in nv04_fifo_runl_ctor() [all...] |
H A D | cgrp.h | 34 struct nvkm_runl *runl; member 71 #define CGRP_PRCLI(c,l,p,f,a...) RUNL_PRINT((c)->runl, l, p, "%04x:[%s]"f, (c)->id, (c)->name, ##a) 72 #define CGRP_PRINT(c,l,p,f,a...) RUNL_PRINT((c)->runl, l, p, "%04x:"f, (c)->id, ##a)
|
H A D | gv100.c | 26 #include "runl.h" 177 gv100_runl_preempt(struct nvkm_runl *runl) in gv100_runl_preempt() argument 179 nvkm_wr32(runl->fifo->engine.subdev.device, 0x002638, BIT(runl->id)); in gv100_runl_preempt() 458 struct nvkm_runl *runl; in gv100_fifo_intr_ctxsw_timeout() local 461 nvkm_runl_foreach(runl, fifo) { in gv100_fifo_intr_ctxsw_timeout() 462 nvkm_runl_foreach_engn_cond(engn, runl, engm & BIT(engn->id)) in gv100_fifo_intr_ctxsw_timeout() 463 nvkm_runl_rc_engn(runl, engn); in gv100_fifo_intr_ctxsw_timeout() 479 .runl = &gv100_runl,
|
H A D | nv40.c | 28 #include "runl.h" 41 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv40_chan_ramfc_write() 113 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv40_eobj_ramht_add() 127 struct nvkm_fifo *fifo = chan->cgrp->runl->fifo; in nv40_ectx_bind() 240 .runl = &nv04_runl,
|
H A D | nv10.c | 27 #include "runl.h" 39 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv10_chan_ramfc_write() 101 .runl = &nv04_runl,
|
H A D | nv17.c | 28 #include "runl.h" 40 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv17_chan_ramfc_write() 130 .runl = &nv04_runl,
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | gk104.c | 152 gk104_fifo_runlist_commit(struct gk104_fifo *fifo, int runl, in gk104_fifo_runlist_commit() argument 169 nvkm_wr32(device, 0x002274, (runl << 20) | nr); in gk104_fifo_runlist_commit() 172 if (!(nvkm_rd32(device, 0x002284 + (runl * 0x08)) & 0x00100000)) in gk104_fifo_runlist_commit() 175 nvkm_error(subdev, "runlist %d update timeout\n", runl); in gk104_fifo_runlist_commit() 179 gk104_fifo_runlist_update(struct gk104_fifo *fifo, int runl) in gk104_fifo_runlist_update() argument 189 mem = fifo->runlist[runl].mem[fifo->runlist[runl].next]; in gk104_fifo_runlist_update() 190 fifo->runlist[runl].next = !fifo->runlist[runl].next; in gk104_fifo_runlist_update() 193 list_for_each_entry(chan, &fifo->runlist[runl] in gk104_fifo_runlist_update() 282 int engn, runl; gk104_fifo_recover_work() local 310 gk104_fifo_recover_runl(struct gk104_fifo *fifo, int runl) gk104_fifo_recover_runl() argument 330 gk104_fifo_recover_chid(struct gk104_fifo *fifo, int runl, int chid) gk104_fifo_recover_chid() argument 362 const u32 runl = (stat & 0x000f0000) >> 16; gk104_fifo_recover_chan() local 401 const u32 runl = fifo->engine[engn].runl; gk104_fifo_recover_engn() local 759 int runl = __ffs(mask); gk104_fifo_intr_runlist() local 884 int runl = mthd - NV_DEVICE_FIFO_RUNLIST_ENGINES(0), engn; gk104_fifo_info() local 909 int engn, runl, pbid, ret, i, j; gk104_fifo_oneinit() local [all...] |
H A D | tu102.c | 32 tu102_fifo_runlist_commit(struct gk104_fifo *fifo, int runl, in tu102_fifo_runlist_commit() argument 39 nvkm_wr32(device, 0x002b00 + (runl * 0x10), lower_32_bits(addr)); in tu102_fifo_runlist_commit() 40 nvkm_wr32(device, 0x002b04 + (runl * 0x10), upper_32_bits(addr)); in tu102_fifo_runlist_commit() 41 nvkm_wr32(device, 0x002b08 + (runl * 0x10), nr); in tu102_fifo_runlist_commit()
|
H A D | gk104.h | 26 int runl; member 72 void (*commit)(struct gk104_fifo *, int runl, 94 void gk104_fifo_runlist_update(struct gk104_fifo *, int runl); 107 void gk104_fifo_runlist_commit(struct gk104_fifo *, int runl,
|
H A D | gpfifogv100.c | 48 nvkm_mask(device, 0x002630, BIT(chan->runl), BIT(chan->runl)); in gv100_fifo_gpfifo_engine_valid() 60 nvkm_mask(device, 0x002630, BIT(chan->runl), 0); in gv100_fifo_gpfifo_engine_valid() 152 chan->runl = runlist; in gv100_fifo_gpfifo_new_()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/ |
H A D | nouveau_chan.c | 525 kfree(drm->runl); in nouveau_channels_fini() 554 drm->runl = kcalloc(drm->runl_nr, sizeof(*drm->runl), GFP_KERNEL); in nouveau_channels_init() 555 if (!drm->runl) in nouveau_channels_init() 570 drm->runl[i].chan_nr = args.v.channels.data; in nouveau_channels_init() 571 drm->runl[i].chan_id_base = drm->chan_total; in nouveau_channels_init() 572 drm->runl[i].context_base = dma_fence_context_alloc(drm->runl[i].chan_nr); in nouveau_channels_init() 574 drm->chan_total += drm->runl[i].chan_nr; in nouveau_channels_init() 577 drm->runl[ in nouveau_channels_init() [all...] |