/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem_ww.c | 9 void i915_gem_ww_ctx_init(struct i915_gem_ww_ctx *ww, bool intr) argument 11 ww_acquire_init(&ww->ctx, &reservation_ww_class); 12 INIT_LIST_HEAD(&ww->obj_list); 13 ww->intr = intr; 14 ww->contended = NULL; 17 static void i915_gem_ww_ctx_unlock_all(struct i915_gem_ww_ctx *ww) argument 21 while ((obj = list_first_entry_or_null(&ww->obj_list, struct drm_i915_gem_object, obj_link))) { 35 void i915_gem_ww_ctx_fini(struct i915_gem_ww_ctx *ww) argument 37 i915_gem_ww_ctx_unlock_all(ww); 38 WARN_ON(ww 42 i915_gem_ww_ctx_backoff(struct i915_gem_ww_ctx *ww) argument [all...] |
H A D | i915_gem_evict.h | 17 struct i915_gem_ww_ctx *ww, 23 struct i915_gem_ww_ctx *ww, 27 struct i915_gem_ww_ctx *ww,
|
H A D | i915_gem_ww.h | 23 static inline int __i915_gem_ww_fini(struct i915_gem_ww_ctx *ww, int err) argument 26 err = i915_gem_ww_ctx_backoff(ww); 32 i915_gem_ww_ctx_fini(ww);
|
H A D | i915_gem_evict.c | 67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument 74 if (!i915_gem_object_trylock(vma->obj, ww)) { 97 struct i915_gem_ww_ctx *ww, 105 if (!grab_vma(vma, ww)) 126 * @ww: An optional struct i915_gem_ww_ctx. 149 struct i915_gem_ww_ctx *ww, 230 if (mark_free(&scan, ww, vma, flags, &eviction_list)) 305 grab_vma(vma, ww)) { 319 * @ww: An optional struct i915_gem_ww_ctx. 329 struct i915_gem_ww_ctx *ww, 96 mark_free(struct drm_mm_scan *scan, struct i915_gem_ww_ctx *ww, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) argument 148 i915_gem_evict_something(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, u64 min_size, u64 alignment, unsigned long color, u64 start, u64 end, unsigned flags) argument 328 i915_gem_evict_for_node(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_mm_node *target, unsigned int flags) argument 458 i915_gem_evict_vm(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_i915_gem_object **busy_bo) argument [all...] |
H A D | i915_gem_gtt.h | 29 struct i915_gem_ww_ctx *ww, 35 struct i915_gem_ww_ctx *ww,
|
H A D | i915_vma.h | 289 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww, 295 struct i915_gem_ww_ctx ww; local 298 i915_gem_ww_ctx_init(&ww, true); 300 err = i915_gem_object_lock(vma->obj, &ww); 302 err = i915_vma_pin_ww(vma, &ww, size, alignment, flags); 304 err = i915_gem_ww_ctx_backoff(&ww); 308 i915_gem_ww_ctx_fini(&ww); 313 int i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
|
/linux-master/drivers/staging/vt6655/ |
H A D | mac.c | 268 unsigned short ww; local 273 for (ww = 0; ww < W_MAX_TIMEOUT; ww++) { 277 if (ww == W_MAX_TIMEOUT) 328 unsigned short ww; local 335 for (ww = 0; ww < W_MAX_TIMEOUT; ww++) { 339 if (ww 382 unsigned short ww; local 533 unsigned short ww; local 567 unsigned short ww; local 601 unsigned short ww; local 636 unsigned short ww; local 737 unsigned int ww; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shrinker.h | 16 unsigned long i915_gem_shrink(struct i915_gem_ww_ctx *ww,
|
H A D | i915_gem_dmabuf.c | 122 struct i915_gem_ww_ctx ww; local 125 i915_gem_ww_ctx_init(&ww, true); 127 err = i915_gem_object_lock(obj, &ww); 135 err = i915_gem_ww_ctx_backoff(&ww); 139 i915_gem_ww_ctx_fini(&ww); 146 struct i915_gem_ww_ctx ww; local 149 i915_gem_ww_ctx_init(&ww, true); 151 err = i915_gem_object_lock(obj, &ww); 159 err = i915_gem_ww_ctx_backoff(&ww); 163 i915_gem_ww_ctx_fini(&ww); 171 struct i915_gem_ww_ctx ww; local [all...] |
H A D | i915_gem_region.c | 163 * In the future, a non-NULL apply->ww could mean the caller is 166 GEM_WARN_ON(apply->ww); 171 struct i915_gem_ww_ctx ww; local 188 apply->ww = &ww; 189 for_i915_gem_ww(&ww, ret, apply->interruptible) { 190 ret = i915_gem_object_lock(obj, apply->ww);
|
H A D | i915_gem_region.h | 27 * Note that if this function is part of a ww transaction, and 39 * @ww: Locking context used for the transaction. 46 struct i915_gem_ww_ctx *ww; member in struct:i915_gem_apply_to_region
|
H A D | i915_gem_object.h | 164 struct i915_gem_ww_ctx *ww, 170 ret = dma_resv_lock_interruptible(obj->base.resv, ww ? &ww->ctx : NULL); 172 ret = dma_resv_lock(obj->base.resv, ww ? &ww->ctx : NULL); 174 if (!ret && ww) { 176 list_add_tail(&obj->obj_link, &ww->obj_list); 183 ww->contended = obj; 190 struct i915_gem_ww_ctx *ww) 192 return __i915_gem_object_lock(obj, ww, w 163 __i915_gem_object_lock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww, bool intr) argument 189 i915_gem_object_lock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument 195 i915_gem_object_lock_interruptible(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument 202 i915_gem_object_trylock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fb_pin.c | 29 struct i915_gem_ww_ctx ww; local 48 for_i915_gem_ww(&ww, ret, true) { 49 ret = i915_gem_object_lock(obj, &ww); 64 ret = __i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0, 86 ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL); 117 struct i915_gem_ww_ctx ww; local 164 i915_gem_ww_ctx_init(&ww, true); 166 ret = i915_gem_object_lock(obj, &ww); 170 ret = i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0); 176 vma = i915_gem_object_pin_to_display_plane(obj, &ww, alignmen [all...] |
H A D | intel_fbdev_fb.c | 72 struct i915_gem_ww_ctx ww; local 94 for_i915_gem_ww(&ww, ret, false) { 95 ret = i915_gem_object_lock(vma->obj, &ww);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_context.c | 116 static int __context_pin_state(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument 121 err = i915_ggtt_pin(vma, ww, 0, bias | PIN_HIGH); 151 struct i915_gem_ww_ctx *ww) 155 err = intel_ring_pin(ring, ww); 177 struct i915_gem_ww_ctx *ww) 183 err = __ring_active(ce->ring, ww); 187 err = intel_timeline_pin(ce->timeline, ww); 194 err = __context_pin_state(ce->state, ww); 218 struct i915_gem_ww_ctx *ww) 236 err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww); 150 __ring_active(struct intel_ring *ring, struct i915_gem_ww_ctx *ww) argument 176 intel_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument 217 __intel_context_do_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument 312 struct i915_gem_ww_ctx ww; local 509 struct i915_gem_ww_ctx ww; local [all...] |
H A D | intel_migrate.h | 25 struct i915_gem_ww_ctx *ww, 47 struct i915_gem_ww_ctx *ww,
|
H A D | intel_renderstate.c | 167 i915_gem_ww_ctx_init(&so->ww, true); 169 err = intel_context_pin_ww(ce, &so->ww); 177 err = i915_gem_object_lock(so->vma->obj, &so->ww); 181 err = i915_vma_pin_ww(so->vma, &so->ww, 0, 0, PIN_GLOBAL | PIN_HIGH); 197 err = i915_gem_ww_ctx_backoff(&so->ww); 201 i915_gem_ww_ctx_fini(&so->ww); 248 i915_gem_ww_ctx_fini(&so->ww);
|
H A D | intel_renderstate.h | 36 struct i915_gem_ww_ctx ww; member in struct:intel_renderstate
|
/linux-master/kernel/locking/ |
H A D | ww_mutex.h | 177 * Associate the ww_mutex @ww with the context @ww_ctx under which we acquired 181 ww_mutex_lock_acquired(struct ww_mutex *ww, struct ww_acquire_ctx *ww_ctx) argument 190 DEBUG_LOCKS_WARN_ON(ww->ctx); 202 DEBUG_LOCKS_WARN_ON(ww_ctx->contending_lock != ww); 215 DEBUG_LOCKS_WARN_ON(ww_ctx->ww_class != ww->ww_class); 218 ww->ctx = ww_ctx; 392 * [W] ww->ctx = ctx [W] MUTEX_FLAG_WAITERS 394 * [R] MUTEX_FLAG_WAITERS [R] ww->ctx 397 * __ww_mutex_add_waiter() and makes sure we either observe ww->ctx 417 struct ww_mutex *ww; local 444 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); local 545 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); local [all...] |
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | igt_spinner.h | 32 struct i915_gem_ww_ctx *ww);
|
H A D | i915_gem.c | 213 struct i915_gem_ww_ctx ww; local 226 i915_gem_ww_ctx_init(&ww, true); 229 err = i915_gem_object_lock(obj, &ww); 231 err = i915_gem_object_lock_interruptible(obj, &ww); 233 err = i915_gem_object_lock_interruptible(obj2, &ww); 235 err = i915_gem_object_lock(obj2, &ww); 238 err = i915_gem_ww_ctx_backoff(&ww); 242 i915_gem_ww_ctx_fini(&ww);
|
H A D | igt_spinner.c | 43 struct i915_gem_ww_ctx *ww, 54 ret = i915_gem_object_lock(obj, ww); 60 if (!ww) 66 if (ww) 67 ret = i915_vma_pin_ww(*vma, ww, 0, 0, PIN_USER); 81 struct i915_gem_ww_ctx *ww) 90 vaddr = igt_spinner_pin_obj(ce, ww, spin->hws, I915_MAP_WB, &spin->hws_vma); 101 vaddr = igt_spinner_pin_obj(ce, ww, spin->obj, mode, &spin->batch_vma); 42 igt_spinner_pin_obj(struct intel_context *ce, struct i915_gem_ww_ctx *ww, struct drm_i915_gem_object *obj, unsigned int mode, struct i915_vma **vma) argument 79 igt_spinner_pin(struct igt_spinner *spin, struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_migrate.c | 53 struct i915_gem_ww_ctx ww; local 64 for_i915_gem_ww(&ww, err, true) { 65 err = i915_gem_object_lock(obj, &ww); 73 err = i915_gem_object_migrate(obj, &ww, dst); 111 static int lmem_pages_migrate_one(struct i915_gem_ww_ctx *ww, argument 118 err = i915_gem_object_lock(obj, ww); 123 err = i915_vma_pin_ww(vma, ww, obj->base.size, 0, 141 err = i915_gem_object_migrate(obj, ww, INTEL_REGION_SMEM); 160 err = i915_gem_object_migrate(obj, ww, INTEL_REGION_LMEM_0); 192 struct i915_gem_ww_ctx ww; local [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_gsc_uc_heci_cmd_submit.c | 142 struct i915_gem_ww_ctx ww; local 146 i915_gem_ww_ctx_init(&ww, false); 148 err = i915_gem_object_lock(pkt->bb_vma->obj, &ww); 151 err = i915_gem_object_lock(pkt->heci_pkt_vma->obj, &ww); 154 err = intel_context_pin_ww(ce, &ww); 216 err = i915_gem_ww_ctx_backoff(&ww); 224 i915_gem_ww_ctx_fini(&ww);
|
/linux-master/drivers/scsi/aic94xx/ |
H A D | aic94xx_reg.c | 108 #define ASD_READ_SW(ww, type, ord) \ 109 static type asd_read_##ww##_##ord(struct asd_ha_struct *asd_ha, \ 113 u32 map_offs = (reg - io_handle->ww##_base) + asd_mem_offs_##ww();\ 117 #define ASD_WRITE_SW(ww, type, ord) \ 118 static void asd_write_##ww##_##ord(struct asd_ha_struct *asd_ha, \ 122 u32 map_offs = (reg - io_handle->ww##_base) + asd_mem_offs_##ww();\
|