/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_device.c | 156 spin_lock(&bdev->lru_lock); 171 /* ttm_bo_swapout has dropped the lru_lock */ 178 spin_unlock(&bdev->lru_lock); 232 spin_lock_init(&bdev->lru_lock); 259 spin_lock(&bdev->lru_lock); 263 spin_unlock(&bdev->lru_lock); 275 spin_lock(&bdev->lru_lock); 279 /* Take ref against racing releases once lru_lock is unlocked */ 284 spin_unlock(&bdev->lru_lock); 290 spin_lock(&bdev->lru_lock); [all...] |
H A D | ttm_resource.c | 53 * resource order never changes. Should be called with &ttm_device.lru_lock held. 67 lockdep_assert_held(&pos->first->bo->bdev->lru_lock); 153 lockdep_assert_held(&bo->bdev->lru_lock); 196 spin_lock(&bo->bdev->lru_lock); 202 spin_unlock(&bo->bdev->lru_lock); 221 spin_lock(&bdev->lru_lock); 224 spin_unlock(&bdev->lru_lock); 240 spin_lock(&bo->bdev->lru_lock); 242 spin_unlock(&bo->bdev->lru_lock); 254 spin_lock(&bo->bdev->lru_lock); [all...] |
H A D | ttm_bo.c | 71 * object. This function must be called with struct ttm_global::lru_lock 105 spin_lock(&bo->bdev->lru_lock); 111 spin_unlock(&bo->bdev->lru_lock); 203 * the resv object while holding the lru_lock. 205 spin_lock(&bo->bdev->lru_lock); 207 spin_unlock(&bo->bdev->lru_lock); 232 * Must be called with lru_lock and reservation held, this function 258 spin_unlock(&bo->bdev->lru_lock); 269 spin_lock(&bo->bdev->lru_lock); 279 spin_unlock(&bo->bdev->lru_lock); [all...] |
/linux-master/mm/ |
H A D | workingset.c | 706 spinlock_t *lru_lock, 707 void *arg) __must_hold(lru_lock) 716 * lru_lock. Because the page cache tree is emptied before 717 * the inode can be destroyed, holding the lru_lock pins any 722 * to reclaim, take the node off-LRU, and drop the lru_lock. 729 spin_unlock_irq(lru_lock); 738 spin_unlock_irq(lru_lock); 747 spin_unlock(lru_lock); 771 spin_lock_irq(lru_lock);
|
H A D | mmzone.c | 80 spin_lock_init(&lruvec->lru_lock);
|
H A D | vmscan.c | 814 * Context: lru_lock must not be held, interrupts must be enabled. 1594 * lruvec->lru_lock is heavily contended. Some of the functions that 1722 * (2) The lru_lock must not be held. 1809 spin_unlock_irq(&lruvec->lru_lock); 1811 spin_lock_irq(&lruvec->lru_lock); 1821 * !lru //skip lru_lock 1835 spin_unlock_irq(&lruvec->lru_lock); 1838 spin_lock_irq(&lruvec->lru_lock); 1857 spin_unlock_irq(&lruvec->lru_lock); 1860 spin_lock_irq(&lruvec->lru_lock); [all...] |
H A D | swap.c | 302 * Hold lruvec->lru_lock is safe here, since 308 spin_lock_irq(&lruvec->lru_lock); 332 spin_unlock_irq(&lruvec->lru_lock);
|
H A D | memcontrol.c | 1375 spin_lock(&lruvec->lru_lock); 1398 spin_lock_irq(&lruvec->lru_lock); 1423 spin_lock_irqsave(&lruvec->lru_lock, *flags); 1436 * This function must be called under lru_lock, just before a page is added
|
/linux-master/include/drm/ttm/ |
H A D | ttm_device.h | 250 * @lru_lock: Protection for the per manager LRU and ddestroy lists. 252 spinlock_t lru_lock; member in struct:ttm_device
|
H A D | ttm_bo.h | 293 spin_lock(&bo->bdev->lru_lock); 295 spin_unlock(&bo->bdev->lru_lock);
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_exec.c | 309 spin_lock(&xe->ttm.lru_lock); 311 spin_unlock(&xe->ttm.lru_lock);
|
H A D | xe_bo.c | 951 * lru_lock, which will immediately be unlocked again since 955 spin_lock(&ttm_bo->bdev->lru_lock); 957 spin_unlock(&ttm_bo->bdev->lru_lock);
|
/linux-master/include/linux/ |
H A D | memcontrol.h | 1380 spin_lock(&pgdat->__lruvec.lru_lock); 1388 spin_lock_irq(&pgdat->__lruvec.lru_lock); 1397 spin_lock_irqsave(&pgdat->__lruvec.lru_lock, *flagsp); 1660 spin_unlock(&lruvec->lru_lock); 1665 spin_unlock_irq(&lruvec->lru_lock); 1671 spin_unlock_irqrestore(&lruvec->lru_lock, flags);
|
H A D | mm_inline.h | 44 lockdep_assert_held(&lruvec->lru_lock);
|
H A D | mmzone.h | 608 /* per lruvec lru_lock for memcg */ 609 spinlock_t lru_lock; member in struct:lruvec
|
/linux-master/fs/gfs2/ |
H A D | glock.c | 72 static DEFINE_SPINLOCK(lru_lock); 221 spin_lock(&lru_lock); 230 spin_unlock(&lru_lock); 238 spin_lock(&lru_lock); 244 spin_unlock(&lru_lock); 1987 * Must be called under the lru_lock, but may drop and retake this 1988 * lock. While the lru_lock is dropped, entries may vanish from the 1994 __releases(&lru_lock) 1995 __acquires(&lru_lock) 2022 cond_resched_lock(&lru_lock); variable [all...] |
H A D | quota.c | 155 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 154 gfs2_qd_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument
|
/linux-master/fs/xfs/ |
H A D | xfs_qm.c | 415 spinlock_t *lru_lock, 417 __releases(lru_lock) __acquires(lru_lock) 463 spin_unlock(lru_lock); 499 spin_lock(lru_lock);
|
H A D | xfs_buf.c | 39 * lru_lock 44 * lru_lock 47 * lru_lock 51 * lru_lock 1857 spinlock_t *lru_lock, 1956 spinlock_t *lru_lock, 1854 xfs_buftarg_drain_rele( struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument 1953 xfs_buftarg_isolate( struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vram_mgr.c | 317 spin_lock(&man->bdev->lru_lock); 319 spin_unlock(&man->bdev->lru_lock);
|
H A D | amdgpu_vm.c | 388 spin_lock(&adev->mman.bdev.lru_lock); 390 spin_unlock(&adev->mman.bdev.lru_lock);
|
/linux-master/fs/nfs/ |
H A D | nfs42xattr.c | 805 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 870 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 804 cache_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument 869 entry_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_ttm.c | 980 spin_lock(&bo->bdev->lru_lock); 1009 spin_unlock(&bo->bdev->lru_lock);
|
/linux-master/fs/ |
H A D | inode.c | 806 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 848 spin_unlock(lru_lock); 859 spin_lock(lru_lock); 805 inode_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument
|
H A D | dcache.c | 1080 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 1161 struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) 1079 dentry_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument 1160 dentry_lru_isolate_shrink(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument
|