Searched refs:ww (Results 1 - 25 of 80) sorted by relevance

1234

/linux-master/drivers/gpu/drm/i915/
H A Di915_gem_ww.c9 void i915_gem_ww_ctx_init(struct i915_gem_ww_ctx *ww, bool intr) argument
11 ww_acquire_init(&ww->ctx, &reservation_ww_class);
12 INIT_LIST_HEAD(&ww->obj_list);
13 ww->intr = intr;
14 ww->contended = NULL;
17 static void i915_gem_ww_ctx_unlock_all(struct i915_gem_ww_ctx *ww) argument
21 while ((obj = list_first_entry_or_null(&ww->obj_list, struct drm_i915_gem_object, obj_link))) {
35 void i915_gem_ww_ctx_fini(struct i915_gem_ww_ctx *ww) argument
37 i915_gem_ww_ctx_unlock_all(ww);
38 WARN_ON(ww
42 i915_gem_ww_ctx_backoff(struct i915_gem_ww_ctx *ww) argument
[all...]
H A Di915_gem_evict.h17 struct i915_gem_ww_ctx *ww,
23 struct i915_gem_ww_ctx *ww,
27 struct i915_gem_ww_ctx *ww,
H A Di915_gem_ww.h23 static inline int __i915_gem_ww_fini(struct i915_gem_ww_ctx *ww, int err) argument
26 err = i915_gem_ww_ctx_backoff(ww);
32 i915_gem_ww_ctx_fini(ww);
H A Di915_gem_evict.c67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument
74 if (!i915_gem_object_trylock(vma->obj, ww)) {
97 struct i915_gem_ww_ctx *ww,
105 if (!grab_vma(vma, ww))
126 * @ww: An optional struct i915_gem_ww_ctx.
149 struct i915_gem_ww_ctx *ww,
230 if (mark_free(&scan, ww, vma, flags, &eviction_list))
305 grab_vma(vma, ww)) {
319 * @ww: An optional struct i915_gem_ww_ctx.
329 struct i915_gem_ww_ctx *ww,
96 mark_free(struct drm_mm_scan *scan, struct i915_gem_ww_ctx *ww, struct i915_vma *vma, unsigned int flags, struct list_head *unwind) argument
148 i915_gem_evict_something(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, u64 min_size, u64 alignment, unsigned long color, u64 start, u64 end, unsigned flags) argument
328 i915_gem_evict_for_node(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_mm_node *target, unsigned int flags) argument
458 i915_gem_evict_vm(struct i915_address_space *vm, struct i915_gem_ww_ctx *ww, struct drm_i915_gem_object **busy_bo) argument
[all...]
H A Di915_gem_gtt.h29 struct i915_gem_ww_ctx *ww,
35 struct i915_gem_ww_ctx *ww,
H A Di915_vma.h289 i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
295 struct i915_gem_ww_ctx ww; local
298 i915_gem_ww_ctx_init(&ww, true);
300 err = i915_gem_object_lock(vma->obj, &ww);
302 err = i915_vma_pin_ww(vma, &ww, size, alignment, flags);
304 err = i915_gem_ww_ctx_backoff(&ww);
308 i915_gem_ww_ctx_fini(&ww);
313 int i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
/linux-master/drivers/staging/vt6655/
H A Dmac.c268 unsigned short ww; local
273 for (ww = 0; ww < W_MAX_TIMEOUT; ww++) {
277 if (ww == W_MAX_TIMEOUT)
328 unsigned short ww; local
335 for (ww = 0; ww < W_MAX_TIMEOUT; ww++) {
339 if (ww
382 unsigned short ww; local
533 unsigned short ww; local
567 unsigned short ww; local
601 unsigned short ww; local
636 unsigned short ww; local
737 unsigned int ww; local
[all...]
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_shrinker.h16 unsigned long i915_gem_shrink(struct i915_gem_ww_ctx *ww,
H A Di915_gem_dmabuf.c122 struct i915_gem_ww_ctx ww; local
125 i915_gem_ww_ctx_init(&ww, true);
127 err = i915_gem_object_lock(obj, &ww);
135 err = i915_gem_ww_ctx_backoff(&ww);
139 i915_gem_ww_ctx_fini(&ww);
146 struct i915_gem_ww_ctx ww; local
149 i915_gem_ww_ctx_init(&ww, true);
151 err = i915_gem_object_lock(obj, &ww);
159 err = i915_gem_ww_ctx_backoff(&ww);
163 i915_gem_ww_ctx_fini(&ww);
171 struct i915_gem_ww_ctx ww; local
[all...]
H A Di915_gem_region.c163 * In the future, a non-NULL apply->ww could mean the caller is
166 GEM_WARN_ON(apply->ww);
171 struct i915_gem_ww_ctx ww; local
188 apply->ww = &ww;
189 for_i915_gem_ww(&ww, ret, apply->interruptible) {
190 ret = i915_gem_object_lock(obj, apply->ww);
H A Di915_gem_region.h27 * Note that if this function is part of a ww transaction, and
39 * @ww: Locking context used for the transaction.
46 struct i915_gem_ww_ctx *ww; member in struct:i915_gem_apply_to_region
H A Di915_gem_object.h164 struct i915_gem_ww_ctx *ww,
170 ret = dma_resv_lock_interruptible(obj->base.resv, ww ? &ww->ctx : NULL);
172 ret = dma_resv_lock(obj->base.resv, ww ? &ww->ctx : NULL);
174 if (!ret && ww) {
176 list_add_tail(&obj->obj_link, &ww->obj_list);
183 ww->contended = obj;
190 struct i915_gem_ww_ctx *ww)
192 return __i915_gem_object_lock(obj, ww, w
163 __i915_gem_object_lock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww, bool intr) argument
189 i915_gem_object_lock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument
195 i915_gem_object_lock_interruptible(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument
202 i915_gem_object_trylock(struct drm_i915_gem_object *obj, struct i915_gem_ww_ctx *ww) argument
[all...]
/linux-master/drivers/gpu/drm/i915/display/
H A Dintel_fb_pin.c29 struct i915_gem_ww_ctx ww; local
48 for_i915_gem_ww(&ww, ret, true) {
49 ret = i915_gem_object_lock(obj, &ww);
64 ret = __i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0,
86 ret = i915_vma_pin_ww(vma, &ww, 0, alignment, PIN_GLOBAL);
117 struct i915_gem_ww_ctx ww; local
164 i915_gem_ww_ctx_init(&ww, true);
166 ret = i915_gem_object_lock(obj, &ww);
170 ret = i915_gem_object_migrate(obj, &ww, INTEL_REGION_LMEM_0);
176 vma = i915_gem_object_pin_to_display_plane(obj, &ww, alignmen
[all...]
H A Dintel_fbdev_fb.c72 struct i915_gem_ww_ctx ww; local
94 for_i915_gem_ww(&ww, ret, false) {
95 ret = i915_gem_object_lock(vma->obj, &ww);
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_context.c116 static int __context_pin_state(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) argument
121 err = i915_ggtt_pin(vma, ww, 0, bias | PIN_HIGH);
151 struct i915_gem_ww_ctx *ww)
155 err = intel_ring_pin(ring, ww);
177 struct i915_gem_ww_ctx *ww)
183 err = __ring_active(ce->ring, ww);
187 err = intel_timeline_pin(ce->timeline, ww);
194 err = __context_pin_state(ce->state, ww);
218 struct i915_gem_ww_ctx *ww)
236 err = i915_gem_object_lock(ce->timeline->hwsp_ggtt->obj, ww);
150 __ring_active(struct intel_ring *ring, struct i915_gem_ww_ctx *ww) argument
176 intel_context_pre_pin(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
217 __intel_context_do_pin_ww(struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
312 struct i915_gem_ww_ctx ww; local
509 struct i915_gem_ww_ctx ww; local
[all...]
H A Dintel_migrate.h25 struct i915_gem_ww_ctx *ww,
47 struct i915_gem_ww_ctx *ww,
H A Dintel_renderstate.c167 i915_gem_ww_ctx_init(&so->ww, true);
169 err = intel_context_pin_ww(ce, &so->ww);
177 err = i915_gem_object_lock(so->vma->obj, &so->ww);
181 err = i915_vma_pin_ww(so->vma, &so->ww, 0, 0, PIN_GLOBAL | PIN_HIGH);
197 err = i915_gem_ww_ctx_backoff(&so->ww);
201 i915_gem_ww_ctx_fini(&so->ww);
248 i915_gem_ww_ctx_fini(&so->ww);
H A Dintel_renderstate.h36 struct i915_gem_ww_ctx ww; member in struct:intel_renderstate
/linux-master/kernel/locking/
H A Dww_mutex.h177 * Associate the ww_mutex @ww with the context @ww_ctx under which we acquired
181 ww_mutex_lock_acquired(struct ww_mutex *ww, struct ww_acquire_ctx *ww_ctx) argument
190 DEBUG_LOCKS_WARN_ON(ww->ctx);
202 DEBUG_LOCKS_WARN_ON(ww_ctx->contending_lock != ww);
215 DEBUG_LOCKS_WARN_ON(ww_ctx->ww_class != ww->ww_class);
218 ww->ctx = ww_ctx;
392 * [W] ww->ctx = ctx [W] MUTEX_FLAG_WAITERS
394 * [R] MUTEX_FLAG_WAITERS [R] ww->ctx
397 * __ww_mutex_add_waiter() and makes sure we either observe ww->ctx
417 struct ww_mutex *ww; local
444 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); local
545 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); local
[all...]
/linux-master/drivers/gpu/drm/i915/selftests/
H A Digt_spinner.h32 struct i915_gem_ww_ctx *ww);
H A Di915_gem.c213 struct i915_gem_ww_ctx ww; local
226 i915_gem_ww_ctx_init(&ww, true);
229 err = i915_gem_object_lock(obj, &ww);
231 err = i915_gem_object_lock_interruptible(obj, &ww);
233 err = i915_gem_object_lock_interruptible(obj2, &ww);
235 err = i915_gem_object_lock(obj2, &ww);
238 err = i915_gem_ww_ctx_backoff(&ww);
242 i915_gem_ww_ctx_fini(&ww);
H A Digt_spinner.c43 struct i915_gem_ww_ctx *ww,
54 ret = i915_gem_object_lock(obj, ww);
60 if (!ww)
66 if (ww)
67 ret = i915_vma_pin_ww(*vma, ww, 0, 0, PIN_USER);
81 struct i915_gem_ww_ctx *ww)
90 vaddr = igt_spinner_pin_obj(ce, ww, spin->hws, I915_MAP_WB, &spin->hws_vma);
101 vaddr = igt_spinner_pin_obj(ce, ww, spin->obj, mode, &spin->batch_vma);
42 igt_spinner_pin_obj(struct intel_context *ce, struct i915_gem_ww_ctx *ww, struct drm_i915_gem_object *obj, unsigned int mode, struct i915_vma **vma) argument
79 igt_spinner_pin(struct igt_spinner *spin, struct intel_context *ce, struct i915_gem_ww_ctx *ww) argument
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_migrate.c53 struct i915_gem_ww_ctx ww; local
64 for_i915_gem_ww(&ww, err, true) {
65 err = i915_gem_object_lock(obj, &ww);
73 err = i915_gem_object_migrate(obj, &ww, dst);
111 static int lmem_pages_migrate_one(struct i915_gem_ww_ctx *ww, argument
118 err = i915_gem_object_lock(obj, ww);
123 err = i915_vma_pin_ww(vma, ww, obj->base.size, 0,
141 err = i915_gem_object_migrate(obj, ww, INTEL_REGION_SMEM);
160 err = i915_gem_object_migrate(obj, ww, INTEL_REGION_LMEM_0);
192 struct i915_gem_ww_ctx ww; local
[all...]
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_gsc_uc_heci_cmd_submit.c142 struct i915_gem_ww_ctx ww; local
146 i915_gem_ww_ctx_init(&ww, false);
148 err = i915_gem_object_lock(pkt->bb_vma->obj, &ww);
151 err = i915_gem_object_lock(pkt->heci_pkt_vma->obj, &ww);
154 err = intel_context_pin_ww(ce, &ww);
216 err = i915_gem_ww_ctx_backoff(&ww);
224 i915_gem_ww_ctx_fini(&ww);
/linux-master/drivers/scsi/aic94xx/
H A Daic94xx_reg.c108 #define ASD_READ_SW(ww, type, ord) \
109 static type asd_read_##ww##_##ord(struct asd_ha_struct *asd_ha, \
113 u32 map_offs = (reg - io_handle->ww##_base) + asd_mem_offs_##ww();\
117 #define ASD_WRITE_SW(ww, type, ord) \
118 static void asd_write_##ww##_##ord(struct asd_ha_struct *asd_ha, \
122 u32 map_offs = (reg - io_handle->ww##_base) + asd_mem_offs_##ww();\

Completed in 204 milliseconds

1234