/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_ww.c | 14 ww->contended = NULL; 38 WARN_ON(ww->contended); 46 if (WARN_ON(!ww->contended)) 51 ret = dma_resv_lock_slow_interruptible(ww->contended->base.resv, &ww->ctx); 53 dma_resv_lock_slow(ww->contended->base.resv, &ww->ctx); 56 list_add_tail(&ww->contended->obj_link, &ww->obj_list); 58 i915_gem_object_put(ww->contended); 60 ww->contended = NULL;
|
H A D | i915_gem_ww.h | 13 struct drm_i915_gem_object *contended; member in struct:i915_gem_ww_ctx
|
/linux-master/arch/arm/include/asm/ |
H A D | spinlock.h | 83 unsigned long contended, res; local 94 : "=&r" (slock), "=&r" (contended), "=&r" (res) 99 if (!contended) { 160 unsigned long contended, res; local 169 : "=&r" (contended), "=&r" (res) 174 if (!contended) { 250 unsigned long contended, res; local 259 : "=&r" (contended), "=&r" (res) 265 if (contended < 0x80000000) {
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_exec.c | 15 * unlocks all previously locked GEM objects and locks the contended one first 90 exec->contended = DRM_EXEC_DUMMY; 106 if (exec->contended != DRM_EXEC_DUMMY) { 107 drm_gem_object_put(exec->contended); 123 if (likely(!exec->contended)) { 128 if (likely(exec->contended == DRM_EXEC_DUMMY)) { 129 exec->contended = NULL; 162 /* Make sure the contended object is locked first */ 165 struct drm_gem_object *obj = exec->contended; 172 exec->contended [all...] |
H A D | drm_modeset_lock.c | 106 drm_printf(&p, "attempting to lock a contended lock without backoff:\n%s", buf); 276 if (WARN_ON(ctx->contended)) 296 if (WARN_ON(ctx->contended)) 327 ctx->contended = lock; 340 * block until the contended lock becomes available. 348 struct drm_modeset_lock *contended = ctx->contended; local 350 ctx->contended = NULL; 353 if (WARN_ON(!contended)) 358 return modeset_lock(contended, ct [all...] |
H A D | drm_gem.c | 1249 int contended = -1; local 1255 if (contended != -1) { 1256 struct drm_gem_object *obj = objs[contended]; 1267 if (i == contended) 1278 if (contended != -1 && contended >= i) 1279 dma_resv_unlock(objs[contended]->resv); 1282 contended = i;
|
/linux-master/include/drm/ |
H A D | drm_exec.h | 44 * @contended: contended GEM object we backed off for 46 struct drm_gem_object *contended; member in struct:drm_exec 135 return !!exec->contended;
|
H A D | drm_modeset_lock.h | 36 * @contended: used internally for -EDEADLK handling 51 * Contended lock: if a lock is contended you should only call 53 * contended lock. 55 struct drm_modeset_lock *contended; member in struct:drm_modeset_acquire_ctx 58 * Stack depot for debugging when a contended lock was not backed off
|
/linux-master/kernel/locking/ |
H A D | percpu-rwsem.c | 226 bool contended = false; local 241 contended = true; 254 if (contended)
|
H A D | test-ww_mutex.c | 444 int contended = -1; local 451 if (n == contended) 461 if (contended > n) 462 ww_mutex_unlock(&locks[order[contended]]); 463 contended = n; 469 ww_mutex_lock_slow(&locks[order[contended]], &ctx);
|
/linux-master/security/keys/ |
H A D | gc.c | 271 goto contended; 274 contended:
|
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_submit.c | 124 int contended, slow_locked = -1, i, ret = 0; local 133 contended = i; 160 obj = &submit->bos[contended].obj->base; 165 submit->bos[contended].flags |= BO_LOCKED; 166 slow_locked = contended;
|
/linux-master/arch/arm/common/ |
H A D | mcpm_head.S | 168 @ In the contended case, non-first men wait here for cluster setup
|
/linux-master/mm/ |
H A D | compaction.c | 549 * very heavily contended. For async compaction, trylock and record if the 550 * lock is contended. The lock will still be acquired but compaction will 560 /* Track if the lock is contended in async mode */ 561 if (cc->mode == MIGRATE_ASYNC && !cc->contended) { 565 cc->contended = true; 574 * very heavily contended. The lock should be periodically unlocked to avoid 593 cc->contended = true; 975 cc->contended = true; 1272 * or a lock is contended. For contention, isolate quickly to 1276 !cc->finish_pageblock && !cc->contended) { [all...] |
H A D | rmap.c | 587 rwc->contended = true; 996 return rwc.contended ? -1 : pra.referenced; 2550 rwc->contended = true; 2643 rwc->contended = true;
|
H A D | internal.h | 609 bool contended; /* Signal lock contention */ member in struct:compact_control
|
H A D | ksm.c | 3127 rwc->contended = true;
|
/linux-master/include/linux/ |
H A D | rmap.h | 707 * try_lock: bail out if the rmap lock is contended 708 * contended: indicate the rmap traversal bailed out due to lock contention 717 bool contended; member in struct:rmap_walk_control
|
/linux-master/arch/x86/kernel/ |
H A D | hpet.c | 813 goto contended; 828 contended:
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_object.h | 183 ww->contended = obj;
|
/linux-master/drivers/regulator/ |
H A D | core.c | 224 struct regulator_dev *held, *contended; local 239 contended = rdev2; 243 ww_mutex_lock_slow(&contended->mutex, ww_ctx); 244 contended->ref_cnt++; 245 contended->mutex_owner = current; 246 swap(held, contended); 247 ret = regulator_lock_nested(contended, ww_ctx);
|
/linux-master/tools/perf/ |
H A D | builtin-lock.c | 296 DEF_KEY_LOCK(contended, "contended", nr_contended, 10), 307 DEF_KEY_LOCK(contended, "contended", nr_contended, 10), 1352 { "acquire", "acquired", "contended", "release" }; 1881 const char *name[4] = { "acquire", "acquired", "contended", "release" }; 2589 "key for sorting (acquired / contended / avg_wait / wait_total / wait_max / wait_min)"), 2591 "output fields (acquired / contended / avg_wait / wait_total / wait_max / wait_min)"), 2603 "key for sorting (contended / wait_total / wait_max / wait_min / avg_wait)"), 2604 OPT_STRING('F', "field", &output_fields, "contended,wait_tota [all...] |
/linux-master/net/core/ |
H A D | dev.c | 3738 bool contended; local 3783 * Heuristic to force contended enqueues to serialize on a 3792 contended = qdisc_is_running(q) || IS_ENABLED(CONFIG_PREEMPT_RT); 3793 if (unlikely(contended)) 3811 if (unlikely(contended)) { 3813 contended = false; 3825 if (unlikely(contended)) { 3827 contended = false; 3837 if (unlikely(contended))
|