Lines Matching refs:vma_lock
1590 /* Return 1 on zap and vma_lock acquired, 0 on contention (only with @try) */
1597 * vma_lock is nested under mmap_lock for vm_ops callback paths.
1602 * When zapping vmas we need to maintain the mmap_lock => vma_lock
1603 * ordering, which requires using vma_lock to walk vma_list to
1604 * acquire an mm, then dropping vma_lock to get the mmap_lock and
1605 * reacquiring vma_lock. This logic is derived from similar
1612 * vma_lock, thus memory_lock is nested under vma_lock.
1614 * This enables the vm_ops.fault callback to acquire vma_lock,
1622 if (!mutex_trylock(&vdev->vma_lock))
1625 mutex_lock(&vdev->vma_lock);
1641 mutex_unlock(&vdev->vma_lock);
1652 if (!mutex_trylock(&vdev->vma_lock)) {
1658 mutex_lock(&vdev->vma_lock);
1673 mutex_unlock(&vdev->vma_lock);
1683 mutex_unlock(&vdev->vma_lock);
1705 /* Caller holds vma_lock */
1735 mutex_lock(&vdev->vma_lock);
1743 mutex_unlock(&vdev->vma_lock);
1753 mutex_lock(&vdev->vma_lock);
1791 mutex_unlock(&vdev->vma_lock);
2182 mutex_init(&vdev->vma_lock);
2199 mutex_destroy(&vdev->vma_lock);
2479 * therefore we need to zap and hold the vma_lock for each device, and only then
2566 mutex_unlock(&cur_mem->vma_lock);
2593 mutex_unlock(&cur->vma_lock);