Searched refs:ptl (Results 26 - 50 of 62) sorted by relevance

123

/linux-master/drivers/platform/surface/aggregator/
H A Dssh_request_layer.c176 /* rtl/ptl may not be set if we're canceling before submitting. */
258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet);
377 * Try to set ptl and check if this request has already been submitted.
381 * ssh_rtl_cancel_nonpending() may run after we've set the ptl
391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) {
397 * Ensure that we set ptl reference before we continue modifying state.
401 * By setting the ptl reference before we test for "locked", we can
592 * ptl reference must have been set previously, as locked is checked
593 * after setting ptl
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dsubpage_prot.c61 spinlock_t *ptl; local
73 pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
83 pte_unmap_unlock(pte - 1, ptl);
/linux-master/mm/
H A Duserfaultfd.c149 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */
181 spinlock_t *ptl; local
195 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl);
235 pte_unmap_unlock(dst_pte, ptl);
353 spinlock_t *ptl; local
362 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl);
377 pte_unmap_unlock(dst_pte, ptl);
435 spinlock_t *ptl; local
439 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl);
459 pte_unmap_unlock(dst_pte, ptl);
1647 spinlock_t *ptl; local
[all...]
H A Dmlock.c368 spinlock_t *ptl; local
375 ptl = pmd_trans_huge_lock(pmd, vma);
376 if (ptl) {
389 start_pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
418 spin_unlock(ptl);
H A Dgup.c616 spinlock_t *ptl; local
626 ptl = huge_pte_lock(h, vma->vm_mm, ptep);
629 spin_unlock(ptl);
906 spinlock_t *ptl; local
915 ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
1003 pte_unmap_unlock(ptep, ptl);
1006 pte_unmap_unlock(ptep, ptl);
1018 spinlock_t *ptl; local
1032 ptl = pmd_lock(mm, pmd);
1034 spin_unlock(ptl);
1073 spinlock_t *ptl; local
[all...]
H A Dmigrate.c309 spinlock_t *ptl; local
314 ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
328 migration_entry_wait_on_locked(entry, ptl);
331 spin_unlock(ptl);
337 * the pte or the ptl from being freed.
343 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep); local
347 spin_lock(ptl);
351 spin_unlock(ptl);
361 migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl); local
369 spinlock_t *ptl; local
[all...]
H A Dhugetlb.c5652 spinlock_t *ptl; local
5680 ptl = huge_pte_lock(h, mm, ptep);
5682 spin_unlock(ptl);
5691 spin_unlock(ptl);
5713 spin_unlock(ptl);
5725 spin_unlock(ptl);
5762 spin_unlock(ptl);
6007 spin_unlock(vmf->ptl);
6043 spin_lock(vmf->ptl);
6083 spin_lock(vmf->ptl);
6189 spinlock_t *ptl; local
6677 spinlock_t *ptl; local
6914 spinlock_t *ptl; local
7632 spinlock_t *ptl; local
[all...]
H A Ddebug_vm_pgtable.c1321 spinlock_t *ptl = NULL; local
1391 args.ptep = pte_offset_map_lock(args.mm, args.pmdp, args.vaddr, &ptl);
1395 pte_unmap_unlock(args.ptep, ptl);
1397 ptl = pmd_lock(args.mm, args.pmdp);
1402 spin_unlock(ptl);
1404 ptl = pud_lock(args.mm, args.pudp);
1409 spin_unlock(ptl);
H A Dpagewalk.c46 spinlock_t *ptl; local
65 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
68 pte_unmap_unlock(pte, ptl);
H A Dmprotect.c88 spinlock_t *ptl; local
96 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
289 pte_unmap_unlock(pte - 1, ptl);
H A Dmempolicy.c544 spinlock_t *ptl; local
546 ptl = pmd_trans_huge_lock(pmd, vma);
547 if (ptl) {
549 spin_unlock(ptl);
553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
607 pte_unmap_unlock(mapped_pte, ptl);
623 spinlock_t *ptl; local
626 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
653 spin_unlock(ptl);
2746 * Make sure ptl i
[all...]
H A Dmemory-failure.c805 spinlock_t *ptl; local
807 ptl = pmd_trans_huge_lock(pmdp, walk->vma);
808 if (ptl) {
810 spin_unlock(ptl);
815 addr, &ptl);
825 pte_unmap_unlock(mapped_pte, ptl);
/linux-master/arch/s390/mm/
H A Dgmap.c556 spinlock_t *ptl; local
612 ptl = pmd_lock(mm, pmd);
634 spin_unlock(ptl);
690 spinlock_t *ptl; local
704 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl);
707 pte_unmap_unlock(ptep, ptl);
864 * @ptl: pointer to the spinlock pointer
869 spinlock_t **ptl)
878 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl);
913 * @ptl
868 gmap_pte_op_walk(struct gmap *gmap, unsigned long gaddr, spinlock_t **ptl) argument
915 gmap_pte_op_end(pte_t *ptep, spinlock_t *ptl) argument
1028 spinlock_t *ptl; local
1151 spinlock_t *ptl; local
1237 spinlock_t *ptl; local
2142 spinlock_t *ptl; local
2497 spinlock_t *ptl; local
[all...]
/linux-master/drivers/input/touchscreen/
H A Dilitek_ts_i2c.c55 struct ilitek_protocol_info ptl; member in struct:ilitek_ts_data
241 ts->ptl.ver = get_unaligned_be16(outbuf);
242 ts->ptl.ver_major = outbuf[0];
414 if (ts->ptl.ver_major == 0x3 ||
415 ts->ptl.ver == BL_V1_6 ||
416 ts->ptl.ver == BL_V1_7)
/linux-master/arch/powerpc/mm/
H A Dpgtable.c334 spinlock_t *ptl; local
353 pte = pte_offset_map_nolock(mm, pmd, addr, &ptl);
355 assert_spin_locked(ptl);
/linux-master/drivers/virt/acrn/
H A Dmm.c180 spinlock_t *ptl; local
191 &ptep, &ptl);
199 pte_unmap_unlock(ptep, ptl);
/linux-master/arch/x86/xen/
H A Dmmu_pv.c667 spinlock_t *ptl = NULL; local
670 ptl = ptlock_ptr(page_ptdesc(page));
671 spin_lock_nest_lock(ptl, &mm->page_table_lock);
674 return ptl;
679 spinlock_t *ptl = v; local
680 spin_unlock(ptl);
702 spinlock_t *ptl; local
724 ptl = NULL;
726 ptl = xen_pte_lock(page, mm);
732 if (ptl) {
826 spinlock_t *ptl = NULL; local
[all...]
/linux-master/include/linux/
H A Drmap.h680 spinlock_t *ptl; member in struct:page_vma_mapped_walk
709 if (pvmw->ptl)
710 spin_unlock(pvmw->ptl);
H A Dmm.h549 /* These three entries are valid only while holding ptl lock */
554 spinlock_t *ptl; /* Page table lock. member in struct:vm_fault
2739 spinlock_t **ptl);
2741 spinlock_t **ptl)
2744 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl));
2931 return ptdesc->ptl;
2949 return &ptdesc->ptl;
2961 * prep_new_page() initialize page->private (and therefore page->ptl)
2965 * slab code uses page->slab_cache, which share storage with page->ptl
2740 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument
3097 spinlock_t *ptl = pmd_lockptr(mm, pmd); local
3135 spinlock_t *ptl = pud_lockptr(mm, pud); local
[all...]
H A Dhugetlb.h1234 spinlock_t *ptl; local
1236 ptl = huge_pte_lockptr(h, mm, pte);
1237 spin_lock(ptl);
1238 return ptl;
/linux-master/scripts/
H A Dleaking_addresses.pl422 state $ptl = get_kernel_config_option('CONFIG_PGTABLE_LEVELS');
424 if ($ptl == 5) {
/linux-master/arch/x86/mm/
H A Dinit.c814 spinlock_t *ptl; local
841 ptep = get_locked_pte(poking_mm, poking_addr, &ptl);
843 pte_unmap_unlock(ptep, ptl);
/linux-master/drivers/media/platform/allegro-dvt/
H A Dallegro-core.c1724 struct nal_hevc_profile_tier_level *ptl; local
1739 ptl = &vps->profile_tier_level;
1740 ptl->general_profile_idc = nal_hevc_profile(profile);
1741 ptl->general_profile_compatibility_flag[ptl->general_profile_idc] = 1;
1742 ptl->general_tier_flag = nal_hevc_tier(tier);
1743 ptl->general_progressive_source_flag = 1;
1744 ptl->general_frame_only_constraint_flag = 1;
1745 ptl->general_level_idc = nal_hevc_level(level);
1763 struct nal_hevc_profile_tier_level *ptl; local
[all...]
/linux-master/include/linux/surface_aggregator/
H A Dserial_hub.h351 /* -- Packet transport layer (ptl). ----------------------------------------- */
470 * @ptl: Pointer to the packet transport layer. May be %NULL if the packet
492 struct ssh_ptl *ptl; member in struct:ssh_packet
/linux-master/arch/x86/mm/pat/
H A Dmemtype.c955 spinlock_t *ptl; local
957 if (follow_pte(vma, vma->vm_start, &ptep, &ptl))
964 pte_unmap_unlock(ptep, ptl);
970 pte_unmap_unlock(ptep, ptl);

Completed in 966 milliseconds

123