/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_4k.c | 28 unsigned long old_pte, new_pte; local 50 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; 52 new_pte |= _PAGE_DIRTY; 53 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); 59 rflags = htab_convert_pte_flags(new_pte, flags); 121 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; 122 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); 127 *ptep = __pte(new_pte & ~H_PAGE_BUSY);
|
H A D | hash_64k.c | 46 unsigned long old_pte, new_pte, subpg_pte; local 68 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED | H_PAGE_COMBO; 70 new_pte |= _PAGE_DIRTY; 71 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); 76 subpg_pte = new_pte & ~subpg_prot; 109 new_pte &= ~H_PAGE_HASHPTE; 131 *ptep = __pte(new_pte & ~H_PAGE_BUSY); 218 new_pte |= pte_set_hidx(ptep, rpte, subpg_index, slot, PTRS_PER_PTE); 219 new_pte |= H_PAGE_HASHPTE; 224 *ptep = __pte(new_pte 235 unsigned long old_pte, new_pte; local [all...] |
H A D | hugetlbpage.c | 26 unsigned long old_pte, new_pte; local 61 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; 63 new_pte |= _PAGE_DIRTY; 64 } while(!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); 70 rflags = htab_convert_pte_flags(new_pte, flags); 101 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; 117 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, offset); 123 *ptep = __pte(new_pte & ~H_PAGE_BUSY);
|
H A D | radix_pgtable.c | 1488 unsigned long old_pte, new_pte; local 1491 new_pte = old_pte | set; 1493 __radix_pte_update(ptep, _PAGE_INVALID, new_pte);
|
/linux-master/arch/arm/mm/ |
H A D | pgd.c | 36 pte_t *new_pte, *init_pte; local 103 new_pte = pte_alloc_map(mm, new_pmd, 0); 104 if (!new_pte) 121 set_pte_ext(new_pte + 0, init_pte[0], 0); 122 set_pte_ext(new_pte + 1, init_pte[1], 0); 124 pte_unmap(new_pte);
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_book3s_64.h | 424 pte_t old_pte, new_pte = __pte(0); local 442 new_pte = pte_mkyoung(old_pte); 444 new_pte = pte_mkdirty(new_pte); 446 if (pte_xchg(ptep, old_pte, new_pte)) 449 return new_pte;
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | hugetlb.h | 62 pte_t old_pte, pte_t new_pte);
|
/linux-master/arch/arm64/include/asm/ |
H A D | hugetlb.h | 60 pte_t old_pte, pte_t new_pte);
|
H A D | pgtable.h | 1364 pte_t old_pte, pte_t new_pte);
|
/linux-master/arch/riscv/kvm/ |
H A D | mmu.c | 139 gpa_t addr, const pte_t *new_pte) 170 set_pte(ptep, *new_pte); 185 pte_t new_pte; local 216 new_pte = pfn_pte(PFN_DOWN(hpa), prot); 217 new_pte = pte_mkdirty(new_pte); 219 return gstage_set_pte(kvm, level, pcache, gpa, &new_pte); 137 gstage_set_pte(struct kvm *kvm, u32 level, struct kvm_mmu_memory_cache *pcache, gpa_t addr, const pte_t *new_pte) argument
|
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_amd.c | 318 pte_t new_pte; local 358 new_pte = pfn_pte(pfn, new_prot); 359 set_pte_atomic(kpte, new_pte);
|
H A D | init_64.c | 296 static void __set_pte_vaddr(pud_t *pud, unsigned long vaddr, pte_t new_pte) argument 301 set_pte(pte, new_pte); 310 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte) argument 315 __set_pte_vaddr(pud, vaddr, new_pte); 318 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) argument 322 __set_pte_vaddr(pud, vaddr, new_pte);
|
/linux-master/arch/x86/include/asm/ |
H A D | pgtable_64.h | 62 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte); 63 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
|
H A D | pgtable.h | 1311 pte_t old_pte, new_pte; local 1315 new_pte = pte_wrprotect(old_pte); 1316 } while (!try_cmpxchg((long *)&ptep->pte, (long *)&old_pte, *(long *)&new_pte));
|
/linux-master/arch/mips/kvm/ |
H A D | mmu.c | 130 pte_t *new_pte; local 134 new_pte = kvm_mmu_memory_cache_alloc(cache); 135 clear_page(new_pte); 136 pmd_populate_kernel(NULL, pmd, new_pte);
|
/linux-master/mm/ |
H A D | mremap.c | 142 pte_t *old_pte, *new_pte, pte; local 178 new_pte = pte_offset_map_nolock(mm, new_pmd, new_addr, &new_ptl); 179 if (!new_pte) { 190 new_pte++, new_addr += PAGE_SIZE) { 210 set_pte_at(mm, new_addr, new_pte, pte); 218 pte_unmap(new_pte - 1);
|
H A D | swapfile.c | 1764 pte_t *pte, new_pte, old_pte; local 1799 new_pte = swp_entry_to_pte(swp_entry); 1831 new_pte = pte_mkold(mk_pte(page, vma->vm_page_prot)); 1833 new_pte = pte_mksoft_dirty(new_pte); 1835 new_pte = pte_mkuffd_wp(new_pte); 1837 set_pte_at(vma->vm_mm, addr, pte, new_pte);
|
H A D | hugetlb.c | 6221 pte_t new_pte; local 6375 new_pte = make_huge_pte(vma, &folio->page, ((vma->vm_flags & VM_WRITE) 6382 new_pte = huge_pte_mkuffd_wp(new_pte); 6383 set_huge_pte_at(mm, haddr, ptep, new_pte, huge_page_size(h));
|
/linux-master/arch/loongarch/kvm/ |
H A D | mmu.c | 806 kvm_pte_t *ptep, new_pte; local 905 new_pte = kvm_pfn_pte(pfn, __pgprot(prot_bits)); 907 new_pte = kvm_pte_mkhuge(new_pte); 919 kvm_set_pte(ptep, new_pte);
|
/linux-master/arch/x86/mm/pat/ |
H A D | set_memory.c | 850 pte_t new_pte, *tmp; local 983 new_pte = pfn_pte(old_pfn, new_prot); 984 __set_pmd_pte(kpte, address, new_pte); 1609 pte_t new_pte; local 1631 new_pte = pfn_pte(pfn, new_prot); 1636 if (pte_val(old_pte) != pte_val(new_pte)) { 1637 set_pte_atomic(kpte, new_pte);
|