/kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
H A D | i915_gem_ww.c | 14 ww->contended = NULL; in i915_gem_ww_ctx_init() 38 WARN_ON(ww->contended); in i915_gem_ww_ctx_fini() 46 if (WARN_ON(!ww->contended)) in i915_gem_ww_ctx_backoff() 51 ret = dma_resv_lock_slow_interruptible(ww->contended->base.resv, &ww->ctx); in i915_gem_ww_ctx_backoff() 53 dma_resv_lock_slow(ww->contended->base.resv, &ww->ctx); in i915_gem_ww_ctx_backoff() 56 list_add_tail(&ww->contended->obj_link, &ww->obj_list); in i915_gem_ww_ctx_backoff() 58 i915_gem_object_put(ww->contended); in i915_gem_ww_ctx_backoff() 60 ww->contended = NULL; in i915_gem_ww_ctx_backoff()
|
H A D | i915_gem_ww.h | 13 struct drm_i915_gem_object *contended; member
|
/kernel/linux/linux-5.10/arch/arm/include/asm/ |
H A D | spinlock.h | 83 unsigned long contended, res; in arch_spin_trylock() local 94 : "=&r" (slock), "=&r" (contended), "=&r" (res) in arch_spin_trylock() 99 if (!contended) { in arch_spin_trylock() 160 unsigned long contended, res; in arch_write_trylock() local 169 : "=&r" (contended), "=&r" (res) in arch_write_trylock() 174 if (!contended) { in arch_write_trylock() 250 unsigned long contended, res; in arch_read_trylock() local 259 : "=&r" (contended), "=&r" (res) in arch_read_trylock() 265 if (contended < 0x80000000) { in arch_read_trylock()
|
/kernel/linux/linux-6.6/arch/arm/include/asm/ |
H A D | spinlock.h | 83 unsigned long contended, res; in arch_spin_trylock() local 94 : "=&r" (slock), "=&r" (contended), "=&r" (res) in arch_spin_trylock() 99 if (!contended) { in arch_spin_trylock() 160 unsigned long contended, res; in arch_write_trylock() local 169 : "=&r" (contended), "=&r" (res) in arch_write_trylock() 174 if (!contended) { in arch_write_trylock() 250 unsigned long contended, res; in arch_read_trylock() local 259 : "=&r" (contended), "=&r" (res) in arch_read_trylock() 265 if (contended < 0x80000000) { in arch_read_trylock()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_exec.c | 15 * unlocks all previously locked GEM objects and locks the contended one first 83 exec->contended = DRM_EXEC_DUMMY; in drm_exec_init() 99 if (exec->contended != DRM_EXEC_DUMMY) { in drm_exec_fini() 100 drm_gem_object_put(exec->contended); in drm_exec_fini() 116 if (likely(!exec->contended)) { in drm_exec_cleanup() 121 if (likely(exec->contended == DRM_EXEC_DUMMY)) { in drm_exec_cleanup() 122 exec->contended = NULL; in drm_exec_cleanup() 155 /* Make sure the contended object is locked first */ 158 struct drm_gem_object *obj = exec->contended; in drm_exec_lock_contended() 165 exec->contended in drm_exec_lock_contended() [all...] |
H A D | drm_modeset_lock.c | 106 drm_printf(&p, "attempting to lock a contended lock without backoff:\n%s", buf); in __drm_stack_depot_print() 276 if (WARN_ON(ctx->contended)) in drm_modeset_drop_locks() 296 if (WARN_ON(ctx->contended)) in modeset_lock() 327 ctx->contended = lock; in modeset_lock() 340 * block until the contended lock becomes available. 348 struct drm_modeset_lock *contended = ctx->contended; in drm_modeset_backoff() local 350 ctx->contended = NULL; in drm_modeset_backoff() 353 if (WARN_ON(!contended)) in drm_modeset_backoff() 358 return modeset_lock(contended, ct in drm_modeset_backoff() [all...] |
H A D | drm_gem.c | 1249 int contended = -1; in drm_gem_lock_reservations() local 1255 if (contended != -1) { in drm_gem_lock_reservations() 1256 struct drm_gem_object *obj = objs[contended]; in drm_gem_lock_reservations() 1267 if (i == contended) in drm_gem_lock_reservations() 1278 if (contended != -1 && contended >= i) in drm_gem_lock_reservations() 1279 dma_resv_unlock(objs[contended]->resv); in drm_gem_lock_reservations() 1282 contended = i; in drm_gem_lock_reservations()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/ |
H A D | drm_modeset_lock.c | 228 WARN_ON(ctx->contended); in drm_modeset_drop_locks() 246 WARN_ON(ctx->contended); in modeset_lock() 276 ctx->contended = lock; in modeset_lock() 288 * block until the contended lock becomes available. 296 struct drm_modeset_lock *contended = ctx->contended; in drm_modeset_backoff() local 298 ctx->contended = NULL; in drm_modeset_backoff() 300 if (WARN_ON(!contended)) in drm_modeset_backoff() 305 return modeset_lock(contended, ctx, ctx->interruptible, true); in drm_modeset_backoff()
|
H A D | drm_gem.c | 1253 int contended = -1; in drm_gem_lock_reservations() local 1259 if (contended != -1) { in drm_gem_lock_reservations() 1260 struct drm_gem_object *obj = objs[contended]; in drm_gem_lock_reservations() 1271 if (i == contended) in drm_gem_lock_reservations() 1282 if (contended != -1 && contended >= i) in drm_gem_lock_reservations() 1283 dma_resv_unlock(objs[contended]->resv); in drm_gem_lock_reservations() 1286 contended = i; in drm_gem_lock_reservations()
|
/kernel/linux/linux-5.10/arch/csky/include/asm/ |
H A D | spinlock.h | 39 u32 tmp, contended, res; in arch_spin_trylock() local 54 : "=&r" (res), "=&r" (tmp), "=&r" (contended) in arch_spin_trylock() 59 if (!contended) in arch_spin_trylock() 62 return !contended; in arch_spin_trylock()
|
/kernel/linux/linux-6.6/include/drm/ |
H A D | drm_exec.h | 44 * @contended: contended GEM object we backed off for 46 struct drm_gem_object *contended; member 135 return !!exec->contended; in drm_exec_is_contended()
|
H A D | drm_modeset_lock.h | 36 * @contended: used internally for -EDEADLK handling 51 * Contended lock: if a lock is contended you should only call 53 * contended lock. 55 struct drm_modeset_lock *contended; member 58 * Stack depot for debugging when a contended lock was not backed off
|
/kernel/linux/linux-5.10/kernel/locking/ |
H A D | test-ww_mutex.c | 391 int contended = -1; in stress_inorder_work() local 398 if (n == contended) in stress_inorder_work() 408 if (contended > n) in stress_inorder_work() 409 ww_mutex_unlock(&locks[order[contended]]); in stress_inorder_work() 410 contended = n; in stress_inorder_work() 415 ww_mutex_lock_slow(&locks[order[contended]], &ctx); in stress_inorder_work()
|
/kernel/linux/linux-6.6/kernel/locking/ |
H A D | test-ww_mutex.c | 431 int contended = -1; in stress_inorder_work() local 438 if (n == contended) in stress_inorder_work() 448 if (contended > n) in stress_inorder_work() 449 ww_mutex_unlock(&locks[order[contended]]); in stress_inorder_work() 450 contended = n; in stress_inorder_work() 455 ww_mutex_lock_slow(&locks[order[contended]], &ctx); in stress_inorder_work()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_submit.c | 124 int contended, slow_locked = -1, i, ret = 0; in submit_lock_objects() local 133 contended = i; in submit_lock_objects() 160 obj = &submit->bos[contended].obj->base; in submit_lock_objects() 165 submit->bos[contended].flags |= BO_LOCKED; in submit_lock_objects() 166 slow_locked = contended; in submit_lock_objects()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_submit.c | 124 int contended, slow_locked = -1, i, ret = 0; in submit_lock_objects() local 133 contended = i; in submit_lock_objects() 160 obj = &submit->bos[contended].obj->base; in submit_lock_objects() 165 submit->bos[contended].flags |= BO_LOCKED; in submit_lock_objects() 166 slow_locked = contended; in submit_lock_objects()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 1329 ww->contended = NULL; in i915_gem_ww_ctx_init() 1351 WARN_ON(ww->contended); in i915_gem_ww_ctx_fini() 1359 if (WARN_ON(!ww->contended)) in i915_gem_ww_ctx_backoff() 1364 ret = dma_resv_lock_slow_interruptible(ww->contended->base.resv, &ww->ctx); in i915_gem_ww_ctx_backoff() 1366 dma_resv_lock_slow(ww->contended->base.resv, &ww->ctx); in i915_gem_ww_ctx_backoff() 1369 list_add_tail(&ww->contended->obj_link, &ww->obj_list); in i915_gem_ww_ctx_backoff() 1371 ww->contended = NULL; in i915_gem_ww_ctx_backoff()
|
H A D | i915_gem.h | 123 struct drm_i915_gem_object *contended; member
|
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/ |
H A D | msm_gem_submit.c | 173 int contended, slow_locked = -1, i, ret = 0; in submit_lock_objects() local 182 contended = i; in submit_lock_objects() 205 struct msm_gem_object *msm_obj = submit->bos[contended].obj; in submit_lock_objects() 210 submit->bos[contended].flags |= BO_LOCKED; in submit_lock_objects() 211 slow_locked = contended; in submit_lock_objects()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/msm/ |
H A D | msm_gem_submit.c | 283 int contended, slow_locked = -1, i, ret = 0; in submit_lock_objects() local 292 contended = i; in submit_lock_objects() 321 struct drm_gem_object *obj = submit->bos[contended].obj; in submit_lock_objects() 326 submit->bos[contended].flags |= BO_LOCKED; in submit_lock_objects() 327 slow_locked = contended; in submit_lock_objects()
|
/kernel/linux/linux-5.10/include/drm/ |
H A D | drm_modeset_lock.h | 34 * @contended: used internally for -EDEADLK handling 48 * Contended lock: if a lock is contended you should only call 50 * contended lock. 52 struct drm_modeset_lock *contended; member
|
/kernel/linux/linux-5.10/security/keys/ |
H A D | gc.c | 271 goto contended; in key_garbage_collector() 274 contended: in key_garbage_collector()
|
/kernel/linux/linux-6.6/security/keys/ |
H A D | gc.c | 271 goto contended; in key_garbage_collector() 274 contended: in key_garbage_collector()
|
/kernel/linux/linux-6.6/mm/ |
H A D | compaction.c | 526 * very heavily contended. For async compaction, trylock and record if the 527 * lock is contended. The lock will still be acquired but compaction will 537 /* Track if the lock is contended in async mode */ in __acquires() 538 if (cc->mode == MIGRATE_ASYNC && !cc->contended) { in __acquires() 542 cc->contended = true; in __acquires() 551 * very heavily contended. The lock should be periodically unlocked to avoid 570 cc->contended = true; in compact_unlock_should_abort() 921 cc->contended = true; in isolate_migratepages_block() 1181 * or a lock is contended. For contention, isolate quickly to in isolate_migratepages_block() 1185 !cc->finish_pageblock && !cc->contended) { in isolate_migratepages_block() [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | compaction.c | 492 * very heavily contended. For async compaction, trylock and record if the 493 * lock is contended. The lock will still be acquired but compaction will 503 /* Track if the lock is contended in async mode */ in __acquires() 504 if (cc->mode == MIGRATE_ASYNC && !cc->contended) { in __acquires() 508 cc->contended = true; in __acquires() 517 * very heavily contended. The lock should be periodically unlocked to avoid 539 cc->contended = true; in compact_unlock_should_abort() 1025 * or a lock is contended. For contention, isolate quickly to in isolate_migratepages_block() 1029 !cc->rescan && !cc->contended) { in isolate_migratepages_block() 2053 if (cc->contended || fatal_signal_pendin in __compact_finished() [all...] |