Searched refs:ptl (Results 1 - 25 of 60) sorted by last modified time

123

/linux-master/mm/
H A Duserfaultfd.c151 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */
184 spinlock_t *ptl; local
197 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl);
238 pte_unmap_unlock(dst_pte, ptl);
356 spinlock_t *ptl; local
365 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl);
380 pte_unmap_unlock(dst_pte, ptl);
438 spinlock_t *ptl; local
442 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl);
462 pte_unmap_unlock(dst_pte, ptl);
1650 spinlock_t *ptl; local
[all...]
H A Dhugetlb.c5658 spinlock_t *ptl; local
5686 ptl = huge_pte_lock(h, mm, ptep);
5688 spin_unlock(ptl);
5697 spin_unlock(ptl);
5719 spin_unlock(ptl);
5731 spin_unlock(ptl);
5768 spin_unlock(ptl);
5928 struct folio *pagecache_folio, spinlock_t *ptl,
6007 spin_unlock(ptl);
6042 spin_lock(ptl);
5926 hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int flags, struct folio *pagecache_folio, spinlock_t *ptl, struct vm_fault *vmf) argument
6186 spinlock_t *ptl; local
6209 spinlock_t *ptl; local
6436 spinlock_t *ptl; local
6672 spinlock_t *ptl; local
6884 spinlock_t *ptl; local
6980 spinlock_t *ptl; local
7698 spinlock_t *ptl; local
[all...]
H A Dmemory-failure.c803 spinlock_t *ptl; local
805 ptl = pmd_trans_huge_lock(pmdp, walk->vma);
806 if (ptl) {
808 spin_unlock(ptl);
813 addr, &ptl);
823 pte_unmap_unlock(mapped_pte, ptl);
H A Dmadvise.c177 spinlock_t *ptl; local
186 ptep = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
198 pte_unmap_unlock(ptep, ptl);
208 pte_unmap_unlock(ptep, ptl);
334 spinlock_t *ptl; local
352 ptl = pmd_trans_huge_lock(pmd, vma);
353 if (!ptl)
379 spin_unlock(ptl);
411 spin_unlock(ptl);
421 start_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
626 spinlock_t *ptl; local
[all...]
H A Dhuge_memory.c901 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd);
913 spin_unlock(vmf->ptl);
930 spin_unlock(vmf->ptl);
937 spin_unlock(vmf->ptl);
1024 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd);
1029 spin_unlock(vmf->ptl);
1032 spin_unlock(vmf->ptl);
1040 spin_unlock(vmf->ptl);
1043 spin_unlock(vmf->ptl);
1063 spinlock_t *ptl; local
1162 spinlock_t *ptl; local
1805 spinlock_t *ptl; local
1885 spinlock_t *ptl; local
2034 spinlock_t *ptl; local
2309 spinlock_t *ptl; local
2326 spinlock_t *ptl; local
2339 spinlock_t *ptl; local
2373 spinlock_t *ptl; local
2652 spinlock_t *ptl; local
[all...]
H A Dgup.c585 spinlock_t *ptl; local
594 ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
682 pte_unmap_unlock(ptep, ptl);
685 pte_unmap_unlock(ptep, ptl);
697 spinlock_t *ptl; local
708 ptl = pmd_lock(mm, pmd);
710 spin_unlock(ptl);
721 ptl = pmd_lock(mm, pmd);
723 spin_unlock(ptl);
727 spin_unlock(ptl);
749 spinlock_t *ptl; local
[all...]
H A Dmemory.c414 spinlock_t *ptl = pmd_lock(mm, pmd); local
435 spin_unlock(ptl);
1576 spinlock_t *ptl; local
1584 start_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
1669 /* Do the actual TLB flush before dropping ptl */
1674 pte_unmap_unlock(start_pte, ptl);
1679 * entries before releasing the ptl), free the batched
1710 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); local
1716 spin_unlock(ptl);
1969 spinlock_t **ptl)
1968 __get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument
2016 spinlock_t *ptl; local
2268 spinlock_t *ptl; local
2494 spinlock_t *ptl; local
2723 spinlock_t *ptl; local
5851 spinlock_t *ptl; local
5945 spinlock_t *ptl; local
5967 spinlock_t *ptl; local
6012 spinlock_t *ptl; local
6429 spinlock_t *ptl; local
[all...]
H A Dfilemap.c1344 * @ptl: already locked ptl. This function will drop the lock.
1349 * should be called while holding the ptl for the migration entry referencing
1352 * Returns after unlocking the ptl.
1357 void migration_entry_wait_on_locked(swp_entry_t entry, spinlock_t *ptl)
1358 __releases(ptl)
1389 * a valid reference to the page, and it must take the ptl to remove the
1390 * migration entry. So the page is valid until the ptl is dropped.
1392 spin_unlock(ptl);
3217 spin_lock(vmf->ptl);
[all...]
H A Dvmscan.c3335 spinlock_t *ptl; local
3345 pte = pte_offset_map_nolock(args->mm, pmd, start & PMD_MASK, &ptl);
3348 if (!spin_trylock(ptl)) {
3396 pte_unmap_unlock(pte, ptl);
3406 spinlock_t *ptl; local
3430 ptl = pmd_lockptr(args->mm, pmd);
3431 if (!spin_trylock(ptl))
3475 spin_unlock(ptl);
4007 lockdep_assert_held(pvmw->ptl);
4010 if (spin_is_contended(pvmw->ptl))
[all...]
H A Dswapfile.c1763 spinlock_t *ptl; local
1781 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
1841 pte_unmap_unlock(pte, ptl);
H A Dmprotect.c88 spinlock_t *ptl; local
96 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
288 pte_unmap_unlock(pte - 1, ptl);
H A Dmlock.c368 spinlock_t *ptl; local
375 ptl = pmd_trans_huge_lock(pmd, vma);
376 if (ptl) {
389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
418 spin_unlock(ptl);
H A Dmigrate.c309 spinlock_t *ptl; local
314 ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
328 migration_entry_wait_on_locked(entry, ptl);
331 spin_unlock(ptl);
337 * the pte or the ptl from being freed.
343 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep); local
347 spin_lock(ptl);
351 spin_unlock(ptl);
361 migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl); local
369 spinlock_t *ptl; local
[all...]
H A Dmempolicy.c544 spinlock_t *ptl; local
546 ptl = pmd_trans_huge_lock(pmd, vma);
547 if (ptl) {
549 spin_unlock(ptl);
553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
607 pte_unmap_unlock(mapped_pte, ptl);
623 spinlock_t *ptl; local
626 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
654 spin_unlock(ptl);
H A Dmemcontrol.c6240 spinlock_t *ptl; local
6242 ptl = pmd_trans_huge_lock(pmd, vma);
6243 if (ptl) {
6251 spin_unlock(ptl);
6255 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
6261 pte_unmap_unlock(pte - 1, ptl);
6438 spinlock_t *ptl; local
6443 ptl = pmd_trans_huge_lock(pmd, vma);
6444 if (ptl) {
6446 spin_unlock(ptl);
[all...]
H A Dkhugepaged.c689 spinlock_t *ptl,
703 * ptl mostly unnecessary.
705 spin_lock(ptl);
707 spin_unlock(ptl);
717 * ptl mostly unnecessary, but preempt has to
721 spin_lock(ptl);
724 spin_unlock(ptl);
775 * @ptl: lock on raw pages' PTEs
784 spinlock_t *ptl,
811 __collapse_huge_page_copy_succeeded(pte, vma, address, ptl,
686 __collapse_huge_page_copy_succeeded(pte_t *pte, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument
778 __collapse_huge_page_copy(pte_t *pte, struct page *page, pmd_t *pmd, pmd_t orig_pmd, struct vm_area_struct *vma, unsigned long address, spinlock_t *ptl, struct list_head *compound_pagelist) argument
1000 spinlock_t *ptl; local
1263 spinlock_t *ptl; local
1490 spinlock_t *pml = NULL, *ptl; local
1697 spinlock_t *ptl; local
[all...]
/linux-master/arch/s390/mm/
H A Dgmap.c556 spinlock_t *ptl; local
612 ptl = pmd_lock(mm, pmd);
634 spin_unlock(ptl);
690 spinlock_t *ptl; local
704 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl);
707 pte_unmap_unlock(ptep, ptl);
864 * @ptl: pointer to the spinlock pointer
869 spinlock_t **ptl)
878 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl);
913 * @ptl
868 gmap_pte_op_walk(struct gmap *gmap, unsigned long gaddr, spinlock_t **ptl) argument
915 gmap_pte_op_end(pte_t *ptep, spinlock_t *ptl) argument
1028 spinlock_t *ptl; local
1151 spinlock_t *ptl; local
1237 spinlock_t *ptl; local
2142 spinlock_t *ptl; local
2497 spinlock_t *ptl; local
[all...]
/linux-master/fs/proc/
H A Dtask_mmu.c617 spinlock_t *ptl; local
619 ptl = pmd_trans_huge_lock(pmd, vma);
620 if (ptl) {
622 spin_unlock(ptl);
626 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
633 pte_unmap_unlock(pte - 1, ptl);
1163 spinlock_t *ptl; local
1166 ptl = pmd_trans_huge_lock(pmd, vma);
1167 if (ptl) {
1183 spin_unlock(ptl);
1462 spinlock_t *ptl; local
2103 spinlock_t *ptl; local
2154 spinlock_t *ptl; local
2256 spinlock_t *ptl; local
2632 spinlock_t *ptl; local
[all...]
/linux-master/include/linux/
H A Dmm.h548 /* These three entries are valid only while holding ptl lock */
553 spinlock_t *ptl; /* Page table lock. member in struct:vm_fault
2700 spinlock_t **ptl);
2702 spinlock_t **ptl)
2705 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl));
2891 return ptdesc->ptl;
2909 return &ptdesc->ptl;
2921 * prep_new_page() initialize page->private (and therefore page->ptl)
2925 * slab code uses page->slab_cache, which share storage with page->ptl
2701 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument
3057 spinlock_t *ptl = pmd_lockptr(mm, pmd); local
3095 spinlock_t *ptl = pud_lockptr(mm, pud); local
[all...]
H A Dmm_types.h426 * @pmd_huge_pte: Protected by ptdesc->ptl, used for THPs.
432 * @ptl: Lock for the page table.
462 spinlock_t *ptl; member in union:ptdesc::__anon1177
464 spinlock_t ptl; member in union:ptdesc::__anon1177
H A Dhugetlb.h1225 spinlock_t *ptl;
1227 ptl = huge_pte_lockptr(h, mm, pte);
1228 spin_lock(ptl);
1229 return ptl; local
/linux-master/virt/kvm/
H A Dkvm_main.c2901 spinlock_t *ptl; local
2904 r = follow_pte(vma->vm_mm, addr, &ptep, &ptl);
2919 r = follow_pte(vma->vm_mm, addr, &ptep, &ptl);
2956 pte_unmap_unlock(ptep, ptl);
/linux-master/drivers/vfio/
H A Dvfio_iommu_type1.c518 spinlock_t *ptl; local
521 ret = follow_pte(vma->vm_mm, vaddr, &ptep, &ptl);
535 ret = follow_pte(vma->vm_mm, vaddr, &ptep, &ptl);
547 pte_unmap_unlock(ptep, ptl);
/linux-master/drivers/perf/
H A Dxgene_pmu.c436 XGENE_PMU_EVENT_ATTR(pa-cp-ptl-req, 0x06),
448 XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req, 0x16),
449 XGENE_PMU_EVENT_ATTR(pa-ptl-rd-req, 0x17),
/linux-master/arch/x86/xen/
H A Dmmu_pv.c667 spinlock_t *ptl = NULL; local
670 ptl = ptlock_ptr(page_ptdesc(page));
671 spin_lock_nest_lock(ptl, &mm->page_table_lock);
674 return ptl;
679 spinlock_t *ptl = v; local
680 spin_unlock(ptl);
702 spinlock_t *ptl; local
724 ptl = NULL;
726 ptl = xen_pte_lock(page, mm);
732 if (ptl) {
826 spinlock_t *ptl = NULL; local
[all...]

Completed in 442 milliseconds

123