/linux-master/mm/ |
H A D | userfaultfd.c | 151 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */ 184 spinlock_t *ptl; local 197 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); 238 pte_unmap_unlock(dst_pte, ptl); 356 spinlock_t *ptl; local 365 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl); 380 pte_unmap_unlock(dst_pte, ptl); 438 spinlock_t *ptl; local 442 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); 462 pte_unmap_unlock(dst_pte, ptl); 1650 spinlock_t *ptl; local [all...] |
H A D | hugetlb.c | 5658 spinlock_t *ptl; local 5686 ptl = huge_pte_lock(h, mm, ptep); 5688 spin_unlock(ptl); 5697 spin_unlock(ptl); 5719 spin_unlock(ptl); 5731 spin_unlock(ptl); 5768 spin_unlock(ptl); 5928 struct folio *pagecache_folio, spinlock_t *ptl, 6007 spin_unlock(ptl); 6042 spin_lock(ptl); 5926 hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int flags, struct folio *pagecache_folio, spinlock_t *ptl, struct vm_fault *vmf) argument 6186 spinlock_t *ptl; local 6209 spinlock_t *ptl; local 6436 spinlock_t *ptl; local 6672 spinlock_t *ptl; local 6884 spinlock_t *ptl; local 6980 spinlock_t *ptl; local 7698 spinlock_t *ptl; local [all...] |
H A D | memory-failure.c | 803 spinlock_t *ptl; local 805 ptl = pmd_trans_huge_lock(pmdp, walk->vma); 806 if (ptl) { 808 spin_unlock(ptl); 813 addr, &ptl); 823 pte_unmap_unlock(mapped_pte, ptl);
|
H A D | madvise.c | 177 spinlock_t *ptl; local 186 ptep = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 198 pte_unmap_unlock(ptep, ptl); 208 pte_unmap_unlock(ptep, ptl); 334 spinlock_t *ptl; local 352 ptl = pmd_trans_huge_lock(pmd, vma); 353 if (!ptl) 379 spin_unlock(ptl); 411 spin_unlock(ptl); 421 start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 626 spinlock_t *ptl; local [all...] |
H A D | huge_memory.c | 901 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); 913 spin_unlock(vmf->ptl); 930 spin_unlock(vmf->ptl); 937 spin_unlock(vmf->ptl); 1024 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); 1029 spin_unlock(vmf->ptl); 1032 spin_unlock(vmf->ptl); 1040 spin_unlock(vmf->ptl); 1043 spin_unlock(vmf->ptl); 1063 spinlock_t *ptl; local 1162 spinlock_t *ptl; local 1805 spinlock_t *ptl; local 1885 spinlock_t *ptl; local 2034 spinlock_t *ptl; local 2309 spinlock_t *ptl; local 2326 spinlock_t *ptl; local 2339 spinlock_t *ptl; local 2373 spinlock_t *ptl; local 2652 spinlock_t *ptl; local [all...] |
H A D | gup.c | 585 spinlock_t *ptl; local 594 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); 682 pte_unmap_unlock(ptep, ptl); 685 pte_unmap_unlock(ptep, ptl); 697 spinlock_t *ptl; local 708 ptl = pmd_lock(mm, pmd); 710 spin_unlock(ptl); 721 ptl = pmd_lock(mm, pmd); 723 spin_unlock(ptl); 727 spin_unlock(ptl); 749 spinlock_t *ptl; local [all...] |
H A D | memory.c | 414 spinlock_t *ptl = pmd_lock(mm, pmd); local 435 spin_unlock(ptl); 1576 spinlock_t *ptl; local 1584 start_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); 1669 /* Do the actual TLB flush before dropping ptl */ 1674 pte_unmap_unlock(start_pte, ptl); 1679 * entries before releasing the ptl), free the batched 1710 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); local 1716 spin_unlock(ptl); 1969 spinlock_t **ptl) 1968 __get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument 2016 spinlock_t *ptl; local 2268 spinlock_t *ptl; local 2494 spinlock_t *ptl; local 2723 spinlock_t *ptl; local 5851 spinlock_t *ptl; local 5945 spinlock_t *ptl; local 5967 spinlock_t *ptl; local 6012 spinlock_t *ptl; local 6429 spinlock_t *ptl; local [all...] |
H A D | filemap.c | 1344 * @ptl: already locked ptl. This function will drop the lock. 1349 * should be called while holding the ptl for the migration entry referencing 1352 * Returns after unlocking the ptl. 1357 void migration_entry_wait_on_locked(swp_entry_t entry, spinlock_t *ptl) 1358 __releases(ptl) 1389 * a valid reference to the page, and it must take the ptl to remove the 1390 * migration entry. So the page is valid until the ptl is dropped. 1392 spin_unlock(ptl); 3217 spin_lock(vmf->ptl); [all...] |
H A D | vmscan.c | 3335 spinlock_t *ptl; local 3345 pte = pte_offset_map_nolock(args->mm, pmd, start & PMD_MASK, &ptl); 3348 if (!spin_trylock(ptl)) { 3396 pte_unmap_unlock(pte, ptl); 3406 spinlock_t *ptl; local 3430 ptl = pmd_lockptr(args->mm, pmd); 3431 if (!spin_trylock(ptl)) 3475 spin_unlock(ptl); 4007 lockdep_assert_held(pvmw->ptl); 4010 if (spin_is_contended(pvmw->ptl)) [all...] |
H A D | swapfile.c | 1763 spinlock_t *ptl; local 1781 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 1841 pte_unmap_unlock(pte, ptl);
|
H A D | mprotect.c | 88 spinlock_t *ptl; local 96 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 288 pte_unmap_unlock(pte - 1, ptl);
|
H A D | mlock.c | 368 spinlock_t *ptl; local 375 ptl = pmd_trans_huge_lock(pmd, vma); 376 if (ptl) { 389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 418 spin_unlock(ptl);
|
H A D | migrate.c | 309 spinlock_t *ptl; local 314 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); 328 migration_entry_wait_on_locked(entry, ptl); 331 spin_unlock(ptl); 337 * the pte or the ptl from being freed. 343 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep); local 347 spin_lock(ptl); 351 spin_unlock(ptl); 361 migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl); local 369 spinlock_t *ptl; local [all...] |
H A D | mempolicy.c | 544 spinlock_t *ptl; local 546 ptl = pmd_trans_huge_lock(pmd, vma); 547 if (ptl) { 549 spin_unlock(ptl); 553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); 607 pte_unmap_unlock(mapped_pte, ptl); 623 spinlock_t *ptl; local 626 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); 654 spin_unlock(ptl);
|
H A D | memcontrol.c | 6240 spinlock_t *ptl; local 6242 ptl = pmd_trans_huge_lock(pmd, vma); 6243 if (ptl) { 6251 spin_unlock(ptl); 6255 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 6261 pte_unmap_unlock(pte - 1, ptl); 6438 spinlock_t *ptl; local 6443 ptl = pmd_trans_huge_lock(pmd, vma); 6444 if (ptl) { 6446 spin_unlock(ptl); [all...] |
H A D | khugepaged.c | 689 spinlock_t *ptl, 703 * ptl mostly unnecessary. 705 spin_lock(ptl); 707 spin_unlock(ptl); 717 * ptl mostly unnecessary, but preempt has to 721 spin_lock(ptl); 724 spin_unlock(ptl); 775 * @ptl: lock on raw pages' PTEs 784 spinlock_t *ptl, 811 __collapse_huge_page_copy_succeeded(pte, vma, address, ptl, 686 __collapse_huge_page_copy_succeeded(pte_t *pte, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument 778 __collapse_huge_page_copy(pte_t *pte, struct page *page, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument 1000 spinlock_t *ptl; local 1263 spinlock_t *ptl; local 1490 spinlock_t *pml = NULL, *ptl; local 1697 spinlock_t *ptl; local [all...] |
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 556 spinlock_t *ptl; local 612 ptl = pmd_lock(mm, pmd); 634 spin_unlock(ptl); 690 spinlock_t *ptl; local 704 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl); 707 pte_unmap_unlock(ptep, ptl); 864 * @ptl: pointer to the spinlock pointer 869 spinlock_t **ptl) 878 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl); 913 * @ptl 868 gmap_pte_op_walk(struct gmap *gmap, unsigned long gaddr, spinlock_t **ptl) argument 915 gmap_pte_op_end(pte_t *ptep, spinlock_t *ptl) argument 1028 spinlock_t *ptl; local 1151 spinlock_t *ptl; local 1237 spinlock_t *ptl; local 2142 spinlock_t *ptl; local 2497 spinlock_t *ptl; local [all...] |
/linux-master/fs/proc/ |
H A D | task_mmu.c | 617 spinlock_t *ptl; local 619 ptl = pmd_trans_huge_lock(pmd, vma); 620 if (ptl) { 622 spin_unlock(ptl); 626 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 633 pte_unmap_unlock(pte - 1, ptl); 1163 spinlock_t *ptl; local 1166 ptl = pmd_trans_huge_lock(pmd, vma); 1167 if (ptl) { 1183 spin_unlock(ptl); 1462 spinlock_t *ptl; local 2103 spinlock_t *ptl; local 2154 spinlock_t *ptl; local 2256 spinlock_t *ptl; local 2632 spinlock_t *ptl; local [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 548 /* These three entries are valid only while holding ptl lock */ 553 spinlock_t *ptl; /* Page table lock. member in struct:vm_fault 2700 spinlock_t **ptl); 2702 spinlock_t **ptl) 2705 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl)); 2891 return ptdesc->ptl; 2909 return &ptdesc->ptl; 2921 * prep_new_page() initialize page->private (and therefore page->ptl) 2925 * slab code uses page->slab_cache, which share storage with page->ptl 2701 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument 3057 spinlock_t *ptl = pmd_lockptr(mm, pmd); local 3095 spinlock_t *ptl = pud_lockptr(mm, pud); local [all...] |
H A D | mm_types.h | 426 * @pmd_huge_pte: Protected by ptdesc->ptl, used for THPs. 432 * @ptl: Lock for the page table. 462 spinlock_t *ptl; member in union:ptdesc::__anon1177 464 spinlock_t ptl; member in union:ptdesc::__anon1177
|
H A D | hugetlb.h | 1225 spinlock_t *ptl; 1227 ptl = huge_pte_lockptr(h, mm, pte); 1228 spin_lock(ptl); 1229 return ptl; local
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 2901 spinlock_t *ptl; local 2904 r = follow_pte(vma->vm_mm, addr, &ptep, &ptl); 2919 r = follow_pte(vma->vm_mm, addr, &ptep, &ptl); 2956 pte_unmap_unlock(ptep, ptl);
|
/linux-master/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 518 spinlock_t *ptl; local 521 ret = follow_pte(vma->vm_mm, vaddr, &ptep, &ptl); 535 ret = follow_pte(vma->vm_mm, vaddr, &ptep, &ptl); 547 pte_unmap_unlock(ptep, ptl);
|
/linux-master/drivers/perf/ |
H A D | xgene_pmu.c | 436 XGENE_PMU_EVENT_ATTR(pa-cp-ptl-req, 0x06), 448 XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req, 0x16), 449 XGENE_PMU_EVENT_ATTR(pa-ptl-rd-req, 0x17),
|
/linux-master/arch/x86/xen/ |
H A D | mmu_pv.c | 667 spinlock_t *ptl = NULL; local 670 ptl = ptlock_ptr(page_ptdesc(page)); 671 spin_lock_nest_lock(ptl, &mm->page_table_lock); 674 return ptl; 679 spinlock_t *ptl = v; local 680 spin_unlock(ptl); 702 spinlock_t *ptl; local 724 ptl = NULL; 726 ptl = xen_pte_lock(page, mm); 732 if (ptl) { 826 spinlock_t *ptl = NULL; local [all...] |