Lines Matching refs:ptl
690 spinlock_t *ptl,
704 * ptl mostly unnecessary.
706 spin_lock(ptl);
708 spin_unlock(ptl);
718 * ptl mostly unnecessary, but preempt has to
722 spin_lock(ptl);
725 spin_unlock(ptl);
776 * @ptl: lock on raw pages' PTEs
781 unsigned long address, spinlock_t *ptl,
808 __collapse_huge_page_copy_succeeded(pte, vma, address, ptl,
983 spinlock_t *ptl;
995 pte = pte_offset_map_nolock(mm, pmd, address, &ptl);
1008 vmf.ptl = ptl;
1247 spinlock_t *ptl;
1259 pte = pte_offset_map_lock(mm, pmd, address, &ptl);
1400 pte_unmap_unlock(pte, ptl);
1481 spinlock_t *pml = NULL, *ptl;
1536 start_pte = pte_offset_map_lock(mm, pmd, haddr, &ptl);
1567 pte_unmap_unlock(start_pte, ptl);
1574 * pmd_lock covers a wider range than ptl, and (if split from mm's
1575 * page_table_lock) ptl nests inside pml. The less time we hold pml,
1579 * ptl before pgt_pmd is removed, so uffd private needs pml taken now.
1584 start_pte = pte_offset_map_nolock(mm, pmd, haddr, &ptl);
1588 spin_lock(ptl);
1589 else if (ptl != pml)
1590 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING);
1601 * We dropped ptl after the first scan, to do the mmu_notifier:
1626 spin_unlock(ptl);
1637 if (ptl != pml)
1638 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING);
1642 if (ptl != pml)
1643 spin_unlock(ptl);
1665 pte_unmap_unlock(start_pte, ptl);
1666 if (pml && pml != ptl)
1687 spinlock_t *ptl;
1725 ptl = pte_lockptr(mm, pmd);
1726 if (ptl != pml)
1727 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING);
1745 if (ptl != pml)
1746 spin_unlock(ptl);