Searched refs:rcl (Results 1 - 9 of 9) sorted by relevance
/kernel/linux/linux-5.10/drivers/gpu/drm/vc4/ |
H A D | vc4_render_cl.c | 50 struct drm_gem_cma_object *rcl; member 58 *(u8 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u8() 64 *(u16 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u16() 70 *(u32 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u32() 331 setup->rcl = &vc4_bo_create(dev, size, true, VC4_BO_TYPE_RCL)->base; in vc4_create_rcl_bo() 332 if (IS_ERR(setup->rcl)) in vc4_create_rcl_bo() 333 return PTR_ERR(setup->rcl); in vc4_create_rcl_bo() 334 list_add_tail(&to_vc4_bo(&setup->rcl->base)->unref_head, in vc4_create_rcl_bo() 377 exec->ct1ca = setup->rcl->paddr; in vc4_create_rcl_bo() 378 exec->ct1ea = setup->rcl in vc4_create_rcl_bo() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/vc4/ |
H A D | vc4_render_cl.c | 50 struct drm_gem_dma_object *rcl; member 58 *(u8 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u8() 64 *(u16 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u16() 70 *(u32 *)(setup->rcl->vaddr + setup->next_offset) = val; in rcl_u32() 331 setup->rcl = &vc4_bo_create(dev, size, true, VC4_BO_TYPE_RCL)->base; in vc4_create_rcl_bo() 332 if (IS_ERR(setup->rcl)) in vc4_create_rcl_bo() 333 return PTR_ERR(setup->rcl); in vc4_create_rcl_bo() 334 list_add_tail(&to_vc4_bo(&setup->rcl->base)->unref_head, in vc4_create_rcl_bo() 377 exec->ct1ca = setup->rcl->dma_addr; in vc4_create_rcl_bo() 378 exec->ct1ea = setup->rcl in vc4_create_rcl_bo() [all...] |
/kernel/linux/linux-6.6/kernel/rcu/ |
H A D | tasks.h | 516 struct rcu_cblist rcl = RCU_CBLIST_INITIALIZER(rcl); in rcu_tasks_invoke_cbs() local 537 rcu_segcblist_extract_done_cbs(&rtpcp->cblist, &rcl); in rcu_tasks_invoke_cbs() 539 len = rcl.len; in rcu_tasks_invoke_cbs() 540 for (rhp = rcu_cblist_dequeue(&rcl); rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_tasks_invoke_cbs()
|
H A D | tree_nocb.h | 333 struct rcu_cblist rcl; in rcu_nocb_do_flush_bypass() local 357 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass() 360 rcu_segcblist_insert_pend_cbs(&rdp->cblist, &rcl); in rcu_nocb_do_flush_bypass()
|
H A D | tree.c | 2129 struct rcu_cblist rcl = RCU_CBLIST_INITIALIZER(rcl); in rcu_do_batch() local 2167 rcu_segcblist_extract_done_cbs(&rdp->cblist, &rcl); in rcu_do_batch() 2176 rhp = rcu_cblist_dequeue(&rcl); in rcu_do_batch() 2178 for (; rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_do_batch() 2225 trace_rcu_batch_end(rcu_state.name, count, !!rcl.head, need_resched(), in rcu_do_batch() 2229 rcu_segcblist_insert_done_cbs(&rdp->cblist, &rcl); in rcu_do_batch()
|
/kernel/linux/linux-5.10/kernel/rcu/ |
H A D | tree.c | 2441 struct rcu_cblist rcl = RCU_CBLIST_INITIALIZER(rcl); in rcu_do_batch() local 2476 rcu_segcblist_extract_done_cbs(&rdp->cblist, &rcl); in rcu_do_batch() 2483 rhp = rcu_cblist_dequeue(&rcl); in rcu_do_batch() 2484 for (; rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_do_batch() 2500 * Note: The rcl structure counts down from zero. in rcu_do_batch() 2503 if (-rcl.len >= bl && (need_resched() || in rcu_do_batch() 2513 if (likely((-rcl.len & 31) || local_clock() < tlimit)) in rcu_do_batch() 2529 count = -rcl.len; in rcu_do_batch() 2531 trace_rcu_batch_end(rcu_state.name, count, !!rcl in rcu_do_batch() [all...] |
H A D | tree_plugin.h | 1692 struct rcu_cblist rcl; in rcu_nocb_do_flush_bypass() local 1704 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass() 1705 rcu_segcblist_insert_pend_cbs(&rdp->cblist, &rcl); in rcu_nocb_do_flush_bypass()
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | emulate.c | 1066 FASTOP2CL(rcl); variable
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | emulate.c | 1031 FASTOP2CL(rcl); variable
|
Completed in 34 milliseconds