/linux-master/drivers/platform/surface/aggregator/ |
H A D | ssh_request_layer.c | 176 /* rtl/ptl may not be set if we're canceling before submitting. */ 258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet); 377 * Try to set ptl and check if this request has already been submitted. 381 * ssh_rtl_cancel_nonpending() may run after we've set the ptl 391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) { 397 * Ensure that we set ptl reference before we continue modifying state. 401 * By setting the ptl reference before we test for "locked", we can 592 * ptl reference must have been set previously, as locked is checked 593 * after setting ptl [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | subpage_prot.c | 61 spinlock_t *ptl; local 73 pte = pte_offset_map_lock(mm, pmd, addr, &ptl); 83 pte_unmap_unlock(pte - 1, ptl);
|
/linux-master/mm/ |
H A D | userfaultfd.c | 149 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */ 181 spinlock_t *ptl; local 195 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); 235 pte_unmap_unlock(dst_pte, ptl); 353 spinlock_t *ptl; local 362 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl); 377 pte_unmap_unlock(dst_pte, ptl); 435 spinlock_t *ptl; local 439 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); 459 pte_unmap_unlock(dst_pte, ptl); 1647 spinlock_t *ptl; local [all...] |
H A D | mlock.c | 368 spinlock_t *ptl; local 375 ptl = pmd_trans_huge_lock(pmd, vma); 376 if (ptl) { 389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 418 spin_unlock(ptl);
|
H A D | gup.c | 616 spinlock_t *ptl; local 626 ptl = huge_pte_lock(h, vma->vm_mm, ptep); 629 spin_unlock(ptl); 906 spinlock_t *ptl; local 915 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); 1003 pte_unmap_unlock(ptep, ptl); 1006 pte_unmap_unlock(ptep, ptl); 1018 spinlock_t *ptl; local 1032 ptl = pmd_lock(mm, pmd); 1034 spin_unlock(ptl); 1073 spinlock_t *ptl; local [all...] |
H A D | migrate.c | 309 spinlock_t *ptl; local 314 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); 328 migration_entry_wait_on_locked(entry, ptl); 331 spin_unlock(ptl); 337 * the pte or the ptl from being freed. 343 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep); local 347 spin_lock(ptl); 351 spin_unlock(ptl); 361 migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl); local 369 spinlock_t *ptl; local [all...] |
H A D | hugetlb.c | 5652 spinlock_t *ptl; local 5680 ptl = huge_pte_lock(h, mm, ptep); 5682 spin_unlock(ptl); 5691 spin_unlock(ptl); 5713 spin_unlock(ptl); 5725 spin_unlock(ptl); 5762 spin_unlock(ptl); 6007 spin_unlock(vmf->ptl); 6043 spin_lock(vmf->ptl); 6083 spin_lock(vmf->ptl); 6189 spinlock_t *ptl; local 6677 spinlock_t *ptl; local 6914 spinlock_t *ptl; local 7632 spinlock_t *ptl; local [all...] |
H A D | debug_vm_pgtable.c | 1321 spinlock_t *ptl = NULL; local 1391 args.ptep = pte_offset_map_lock(args.mm, args.pmdp, args.vaddr, &ptl); 1395 pte_unmap_unlock(args.ptep, ptl); 1397 ptl = pmd_lock(args.mm, args.pmdp); 1402 spin_unlock(ptl); 1404 ptl = pud_lock(args.mm, args.pudp); 1409 spin_unlock(ptl);
|
H A D | pagewalk.c | 46 spinlock_t *ptl; local 65 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); 68 pte_unmap_unlock(pte, ptl);
|
H A D | mprotect.c | 88 spinlock_t *ptl; local 96 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 289 pte_unmap_unlock(pte - 1, ptl);
|
H A D | mempolicy.c | 544 spinlock_t *ptl; local 546 ptl = pmd_trans_huge_lock(pmd, vma); 547 if (ptl) { 549 spin_unlock(ptl); 553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); 607 pte_unmap_unlock(mapped_pte, ptl); 623 spinlock_t *ptl; local 626 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); 653 spin_unlock(ptl); 2746 * Make sure ptl i [all...] |
H A D | memory-failure.c | 805 spinlock_t *ptl; local 807 ptl = pmd_trans_huge_lock(pmdp, walk->vma); 808 if (ptl) { 810 spin_unlock(ptl); 815 addr, &ptl); 825 pte_unmap_unlock(mapped_pte, ptl);
|
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 556 spinlock_t *ptl; local 612 ptl = pmd_lock(mm, pmd); 634 spin_unlock(ptl); 690 spinlock_t *ptl; local 704 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl); 707 pte_unmap_unlock(ptep, ptl); 864 * @ptl: pointer to the spinlock pointer 869 spinlock_t **ptl) 878 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl); 913 * @ptl 868 gmap_pte_op_walk(struct gmap *gmap, unsigned long gaddr, spinlock_t **ptl) argument 915 gmap_pte_op_end(pte_t *ptep, spinlock_t *ptl) argument 1028 spinlock_t *ptl; local 1151 spinlock_t *ptl; local 1237 spinlock_t *ptl; local 2142 spinlock_t *ptl; local 2497 spinlock_t *ptl; local [all...] |
/linux-master/drivers/input/touchscreen/ |
H A D | ilitek_ts_i2c.c | 55 struct ilitek_protocol_info ptl; member in struct:ilitek_ts_data 241 ts->ptl.ver = get_unaligned_be16(outbuf); 242 ts->ptl.ver_major = outbuf[0]; 414 if (ts->ptl.ver_major == 0x3 || 415 ts->ptl.ver == BL_V1_6 || 416 ts->ptl.ver == BL_V1_7)
|
/linux-master/arch/powerpc/mm/ |
H A D | pgtable.c | 334 spinlock_t *ptl; local 353 pte = pte_offset_map_nolock(mm, pmd, addr, &ptl); 355 assert_spin_locked(ptl);
|
/linux-master/drivers/virt/acrn/ |
H A D | mm.c | 180 spinlock_t *ptl; local 191 &ptep, &ptl); 199 pte_unmap_unlock(ptep, ptl);
|
/linux-master/arch/x86/xen/ |
H A D | mmu_pv.c | 667 spinlock_t *ptl = NULL; local 670 ptl = ptlock_ptr(page_ptdesc(page)); 671 spin_lock_nest_lock(ptl, &mm->page_table_lock); 674 return ptl; 679 spinlock_t *ptl = v; local 680 spin_unlock(ptl); 702 spinlock_t *ptl; local 724 ptl = NULL; 726 ptl = xen_pte_lock(page, mm); 732 if (ptl) { 826 spinlock_t *ptl = NULL; local [all...] |
/linux-master/include/linux/ |
H A D | rmap.h | 680 spinlock_t *ptl; member in struct:page_vma_mapped_walk 709 if (pvmw->ptl) 710 spin_unlock(pvmw->ptl);
|
H A D | mm.h | 549 /* These three entries are valid only while holding ptl lock */ 554 spinlock_t *ptl; /* Page table lock. member in struct:vm_fault 2739 spinlock_t **ptl); 2741 spinlock_t **ptl) 2744 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl)); 2931 return ptdesc->ptl; 2949 return &ptdesc->ptl; 2961 * prep_new_page() initialize page->private (and therefore page->ptl) 2965 * slab code uses page->slab_cache, which share storage with page->ptl 2740 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument 3097 spinlock_t *ptl = pmd_lockptr(mm, pmd); local 3135 spinlock_t *ptl = pud_lockptr(mm, pud); local [all...] |
H A D | hugetlb.h | 1234 spinlock_t *ptl; local 1236 ptl = huge_pte_lockptr(h, mm, pte); 1237 spin_lock(ptl); 1238 return ptl;
|
/linux-master/scripts/ |
H A D | leaking_addresses.pl | 422 state $ptl = get_kernel_config_option('CONFIG_PGTABLE_LEVELS'); 424 if ($ptl == 5) {
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 814 spinlock_t *ptl; local 841 ptep = get_locked_pte(poking_mm, poking_addr, &ptl); 843 pte_unmap_unlock(ptep, ptl);
|
/linux-master/drivers/media/platform/allegro-dvt/ |
H A D | allegro-core.c | 1724 struct nal_hevc_profile_tier_level *ptl; local 1739 ptl = &vps->profile_tier_level; 1740 ptl->general_profile_idc = nal_hevc_profile(profile); 1741 ptl->general_profile_compatibility_flag[ptl->general_profile_idc] = 1; 1742 ptl->general_tier_flag = nal_hevc_tier(tier); 1743 ptl->general_progressive_source_flag = 1; 1744 ptl->general_frame_only_constraint_flag = 1; 1745 ptl->general_level_idc = nal_hevc_level(level); 1763 struct nal_hevc_profile_tier_level *ptl; local [all...] |
/linux-master/include/linux/surface_aggregator/ |
H A D | serial_hub.h | 351 /* -- Packet transport layer (ptl). ----------------------------------------- */ 470 * @ptl: Pointer to the packet transport layer. May be %NULL if the packet 492 struct ssh_ptl *ptl; member in struct:ssh_packet
|
/linux-master/arch/x86/mm/pat/ |
H A D | memtype.c | 955 spinlock_t *ptl; local 957 if (follow_pte(vma, vma->vm_start, &ptep, &ptl)) 964 pte_unmap_unlock(ptep, ptl); 970 pte_unmap_unlock(ptep, ptl);
|