Searched refs:_PAGE_INVALID (Results 1 - 16 of 16) sorted by relevance
/linux-master/arch/s390/mm/ |
H A D | dump_pagetables.c | 114 if (pr & _PAGE_INVALID) { 126 if (st->current_prot & _PAGE_INVALID) 159 if (level == 4 && (val & _PAGE_INVALID)) 160 prot = _PAGE_INVALID; 163 prot = _PAGE_INVALID;
|
H A D | pgtable.c | 94 if (unlikely(pte_val(old) & _PAGE_INVALID)) 113 if (unlikely(pte_val(old) & _PAGE_INVALID)) 118 set_pte(ptep, set_pte_bit(*ptep, __pgprot(_PAGE_INVALID))); 170 if (!mm_uses_skeys(mm) || pte_val(pte) & _PAGE_INVALID) 192 if (!mm_uses_skeys(mm) || pte_val(entry) & _PAGE_INVALID) 194 VM_BUG_ON(!(pte_val(*ptep) & _PAGE_INVALID)); 212 !(pte_val(entry) & _PAGE_INVALID)) { 263 if (pte_val(old) & _PAGE_INVALID) 265 if (pte_val(new) & _PAGE_INVALID) { 598 set_pte(ptep, __pte(_PAGE_INVALID)); [all...] |
H A D | hugetlbpage.c | 55 rste |= move_set_bit(pte_val(pte), _PAGE_INVALID, 109 pteval |= move_set_bit(rste, _SEGMENT_ENTRY_INVALID, _PAGE_INVALID); 118 pteval = _PAGE_INVALID;
|
H A D | pgalloc.c | 147 memset64(table, _PAGE_INVALID, PTRS_PER_PTE); 176 memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE); 241 memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE);
|
H A D | pageattr.c | 107 new = set_pte_bit(new, __pgprot(_PAGE_INVALID)); 440 pte = clear_pte_bit(*ptep, __pgprot(_PAGE_INVALID));
|
H A D | vmem.c | 72 memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
|
H A D | gmap.c | 1168 /* Do *NOT* clear the _PAGE_INVALID bit! */ 1318 if (!table || *table & _PAGE_INVALID) 1339 pgt[i] = _PAGE_INVALID; 2576 ptep_xchg_direct(walk->mm, addr, ptep, __pte(_PAGE_INVALID));
|
/linux-master/arch/powerpc/mm/ptdump/ |
H A D | book3s64.c | 44 .mask = _PAGE_PRESENT | _PAGE_INVALID,
|
/linux-master/arch/s390/include/asm/ |
H A D | pgtable.h | 164 #define _PAGE_INVALID 0x400 /* HW invalid bit */ macro 408 #define PAGE_NONE __pgprot(_PAGE_PRESENT | _PAGE_INVALID | _PAGE_PROTECT) 410 _PAGE_NOEXEC | _PAGE_INVALID | _PAGE_PROTECT) 412 _PAGE_INVALID | _PAGE_PROTECT) 414 _PAGE_NOEXEC | _PAGE_INVALID | _PAGE_PROTECT) 416 _PAGE_INVALID | _PAGE_PROTECT) 794 return pte_val(pte) == _PAGE_INVALID; 904 * young/old accounting is not supported, i.e _PAGE_PROTECT and _PAGE_INVALID 974 set_pte(ptep, __pte(_PAGE_INVALID)); 990 pte = clear_pte_bit(pte, __pgprot(_PAGE_INVALID)); [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hugetlbpage.c | 138 _PAGE_PRESENT, _PAGE_INVALID, 1);
|
H A D | pgtable.c | 173 old_pmd = pmd_hugepage_update(vma->vm_mm, address, pmdp, _PAGE_PRESENT, _PAGE_INVALID); 530 pte_val = pte_update(vma->vm_mm, addr, ptep, _PAGE_PRESENT, _PAGE_INVALID, 0);
|
H A D | radix_pgtable.c | 1490 old_pte = __radix_pte_update(ptep, _PAGE_PRESENT, _PAGE_INVALID); 1493 __radix_pte_update(ptep, _PAGE_INVALID, new_pte);
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | pgtable.h | 52 #define _PAGE_INVALID _RPAGE_SW0 macro 501 * invalid during ptep_set_access_flags. Hence we look for _PAGE_INVALID 507 return (pte_raw(pte) & cpu_to_be64(_PAGE_INVALID | _PAGE_PTE)) == 508 cpu_to_be64(_PAGE_INVALID | _PAGE_PTE); 827 * invalid during a split. Hence we look for _PAGE_INVALID 830 if (pmd_raw(pmd) & cpu_to_be64(_PAGE_PRESENT | _PAGE_INVALID)) 840 * and _PAGE_INVALID is set (see pmd_present, pmdp_invalidate). 846 if ((pmd_raw(pmd) & cpu_to_be64(_PAGE_PRESENT | _PAGE_INVALID)) == 847 cpu_to_be64(_PAGE_INVALID))
|
/linux-master/arch/s390/boot/ |
H A D | vmem.c | 253 memset64((u64 *)pte, _PAGE_INVALID, PTRS_PER_PTE);
|
/linux-master/arch/s390/kernel/ |
H A D | uv.c | 308 if (pte_present(*ptep) && !(pte_val(*ptep) & _PAGE_INVALID) && pte_write(*ptep)) {
|
/linux-master/arch/s390/kvm/ |
H A D | priv.c | 1189 if (ptev & _PAGE_INVALID) {
|
Completed in 175 milliseconds