Lines Matching refs:ptl

416 	spinlock_t *ptl = pmd_lock(mm, pmd);
437 spin_unlock(ptl);
1583 spinlock_t *ptl;
1591 start_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
1677 /* Do the actual TLB flush before dropping ptl */
1682 pte_unmap_unlock(start_pte, ptl);
1687 * entries before releasing the ptl), free the batched
1718 spinlock_t *ptl = pmd_lock(tlb->mm, pmd);
1724 spin_unlock(ptl);
1977 spinlock_t **ptl)
1983 return pte_alloc_map_lock(mm, pmd, addr, ptl);
2024 spinlock_t *ptl;
2030 pte = get_locked_pte(vma->vm_mm, addr, &ptl);
2034 pte_unmap_unlock(pte, ptl);
2276 spinlock_t *ptl;
2278 pte = get_locked_pte(mm, addr, &ptl);
2321 pte_unmap_unlock(pte, ptl);
2502 spinlock_t *ptl;
2505 mapped_pte = pte = pte_alloc_map_lock(mm, pmd, addr, &ptl);
2519 pte_unmap_unlock(mapped_pte, ptl);
2731 spinlock_t *ptl;
2736 pte_alloc_map_lock(mm, pmd, addr, &ptl);
2742 pte_offset_map_lock(mm, pmd, addr, &ptl);
2763 pte_unmap_unlock(mapped_pte, ptl);
2952 spin_lock(vmf->ptl);
2954 spin_unlock(vmf->ptl);
3004 vmf->pte = pte_offset_map_lock(mm, vmf->pmd, addr, &vmf->ptl);
3032 vmf->pte = pte_offset_map_lock(mm, vmf->pmd, addr, &vmf->ptl);
3060 pte_unmap_unlock(vmf->pte, vmf->ptl);
3175 __releases(vmf->ptl)
3198 pte_unmap_unlock(vmf->pte, vmf->ptl);
3257 * without the ptl held.
3327 vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl);
3392 pte_unmap_unlock(vmf->pte, vmf->ptl);
3395 pte_unmap_unlock(vmf->pte, vmf->ptl);
3441 &vmf->ptl);
3450 pte_unmap_unlock(vmf->pte, vmf->ptl);
3468 pte_unmap_unlock(vmf->pte, vmf->ptl);
3484 __releases(vmf->ptl)
3494 pte_unmap_unlock(vmf->pte, vmf->ptl);
3594 __releases(vmf->ptl)
3604 pte_unmap_unlock(vmf->pte, vmf->ptl);
3666 pte_unmap_unlock(vmf->pte, vmf->ptl);
3678 pte_unmap_unlock(vmf->pte, vmf->ptl);
3846 &vmf->ptl);
3851 pte_unmap_unlock(vmf->pte, vmf->ptl);
3881 vmf->address, &vmf->ptl);
3894 pte_unmap_unlock(vmf->pte, vmf->ptl);
3991 vmf->address, &vmf->ptl);
4002 pte_unmap_unlock(vmf->pte, vmf->ptl);
4084 vmf->address, &vmf->ptl);
4155 &vmf->ptl);
4291 pte_unmap_unlock(vmf->pte, vmf->ptl);
4301 pte_unmap_unlock(vmf->pte, vmf->ptl);
4438 vmf->address, &vmf->ptl);
4450 pte_unmap_unlock(vmf->pte, vmf->ptl);
4482 vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl);
4500 pte_unmap_unlock(vmf->pte, vmf->ptl);
4521 pte_unmap_unlock(vmf->pte, vmf->ptl);
4639 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd);
4666 spin_unlock(vmf->ptl);
4779 vmf->address, &vmf->ptl);
4783 /* Re-check under ptl */
4796 pte_unmap_unlock(vmf->pte, vmf->ptl);
5037 vmf->address, &vmf->ptl);
5043 * by holding ptl and checking again. A R/M/W update
5044 * of pte involves: take ptl, clearing the pte so that
5053 pte_unmap_unlock(vmf->pte, vmf->ptl);
5153 spin_lock(vmf->ptl);
5158 pte_unmap_unlock(vmf->pte, vmf->ptl);
5211 pte_unmap_unlock(vmf->pte, vmf->ptl);
5222 vmf->address, &vmf->ptl);
5226 pte_unmap_unlock(vmf->pte, vmf->ptl);
5247 pte_unmap_unlock(vmf->pte, vmf->ptl);
5367 vmf->address, &vmf->ptl);
5388 spin_lock(vmf->ptl);
5420 pte_unmap_unlock(vmf->pte, vmf->ptl);
5924 spinlock_t *ptl;
5929 ptl = pud_lock(mm, pud);
5937 spin_unlock(ptl);
6035 spinlock_t *ptl;
6040 if (follow_pte(vma, addr, &ptep, &ptl))
6043 pte_unmap_unlock(ptep, ptl);
6055 if (follow_pte(vma, addr, &ptep, &ptl))
6059 pte_unmap_unlock(ptep, ptl);
6070 pte_unmap_unlock(ptep, ptl);
6443 page_ptl_cachep = kmem_cache_create("page->ptl", sizeof(spinlock_t), 0,
6449 spinlock_t *ptl;
6451 ptl = kmem_cache_alloc(page_ptl_cachep, GFP_KERNEL);
6452 if (!ptl)
6454 ptdesc->ptl = ptl;
6460 kmem_cache_free(page_ptl_cachep, ptdesc->ptl);