/linux-master/arch/s390/include/asm/ |
H A D | mem_encrypt.h | 7 int set_memory_encrypted(unsigned long vaddr, int numpages); 8 int set_memory_decrypted(unsigned long vaddr, int numpages);
|
/linux-master/arch/parisc/mm/ |
H A D | fixmap.c | 15 unsigned long vaddr = __fix_to_virt(idx); local 16 pgd_t *pgd = pgd_offset_k(vaddr); 17 p4d_t *p4d = p4d_offset(pgd, vaddr); 18 pud_t *pud = pud_offset(p4d, vaddr); 19 pmd_t *pmd = pmd_offset(pud, vaddr); 22 pte = pte_offset_kernel(pmd, vaddr); 23 set_pte_at(&init_mm, vaddr, pte, __mk_pte(phys, PAGE_KERNEL_RWX)); 24 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); 29 unsigned long vaddr local [all...] |
/linux-master/arch/mips/kernel/ |
H A D | crash_dump.c | 9 void *vaddr; local 14 vaddr = kmap_local_pfn(pfn); 15 csize = copy_to_iter(vaddr + offset, csize, iter); 16 kunmap_local(vaddr);
|
/linux-master/arch/x86/mm/ |
H A D | maccess.c | 11 unsigned long vaddr = (unsigned long)unsafe_src; local 17 if (vaddr < TASK_SIZE_MAX + PAGE_SIZE) 25 if (is_vsyscall_vaddr(vaddr)) 36 return __is_canonical_address(vaddr, boot_cpu_data.x86_virt_bits);
|
/linux-master/arch/riscv/kernel/ |
H A D | crash_dump.c | 15 void *vaddr; local 20 vaddr = memremap(__pfn_to_phys(pfn), PAGE_SIZE, MEMREMAP_WB); 21 if (!vaddr) 24 csize = copy_to_iter(vaddr + offset, csize, iter); 26 memunmap(vaddr);
|
/linux-master/arch/arm/kernel/ |
H A D | crash_dump.c | 22 void *vaddr; local 27 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); 28 if (!vaddr) 31 csize = copy_to_iter(vaddr + offset, csize, iter); 33 iounmap(vaddr);
|
/linux-master/arch/loongarch/kernel/ |
H A D | crash_dump.c | 9 void *vaddr; local 14 vaddr = memremap(__pfn_to_phys(pfn), PAGE_SIZE, MEMREMAP_WB); 15 if (!vaddr) 18 csize = copy_to_iter(vaddr + offset, csize, iter); 20 memunmap(vaddr);
|
/linux-master/arch/arm/mm/ |
H A D | cache-xsc3l2.c | 88 unsigned long vaddr; local 95 vaddr = -1; /* to force the first mapping */ 101 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); 102 xsc3_l2_clean_mva(vaddr); 103 xsc3_l2_inv_mva(vaddr); 111 vaddr = l2_map_va(start, vaddr); 112 xsc3_l2_inv_mva(vaddr); 120 vaddr 132 unsigned long vaddr; local 170 unsigned long vaddr; local [all...] |
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | page.h | 3 static inline void clear_user_page(void *addr, unsigned long vaddr, argument 9 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, argument
|
/linux-master/arch/m68k/include/asm/ |
H A D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 54 : "d" (nr ^ 31), "o" (*vaddr) 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) 74 bclr_reg_clear_bit(int nr, volatile unsigned long *vaddr) argument 84 bclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 93 bfclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 117 bchg_reg_change_bit(int nr, volatile unsigned long *vaddr) argument 127 bchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 136 bfchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 163 bset_reg_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 176 bset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 188 bfset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 216 bclr_reg_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 229 bclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 241 bfclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 269 bchg_reg_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 282 bchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 294 bfchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 353 find_first_zero_bit(const unsigned long *vaddr, unsigned size) argument 379 find_next_zero_bit(const unsigned long *vaddr, int size, int offset) argument 409 find_first_bit(const unsigned long *vaddr, unsigned size) argument 434 find_next_bit(const unsigned long *vaddr, int size, int offset) argument [all...] |
H A D | page_no.h | 13 #define clear_user_page(page, vaddr, pg) clear_page(page) 14 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) 16 #define vma_alloc_zeroed_movable_folio(vma, vaddr) \ 17 vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false) 19 #define __pa(vaddr) ((unsigned long)(vaddr))
|
/linux-master/arch/csky/abiv1/inc/abi/ |
H A D | page.h | 13 static inline void clear_user_page(void *addr, unsigned long vaddr, argument 17 if (pages_do_alias((unsigned long) addr, vaddr & PAGE_MASK)) 21 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, argument 25 if (pages_do_alias((unsigned long) to, vaddr & PAGE_MASK))
|
/linux-master/arch/sh/kernel/ |
H A D | crash_dump.c | 17 void __iomem *vaddr; local 22 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 23 csize = copy_to_iter(vaddr + offset, csize, iter); 24 iounmap(vaddr);
|
/linux-master/arch/m68k/sun3/ |
H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) argument 34 // pr_info("dvma_remap: addr %lx -> %lx pte %08lx\n", kaddr, vaddr, pte); 35 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { 36 sun3_put_pte(vaddr, pte); 37 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; 40 return (vaddr + (kaddr & ~PAGE_MASK)); 49 unsigned long vaddr; local 51 vaddr = dvma_btov(baddr); 53 end = vaddr + len; 55 while(vaddr < en [all...] |
/linux-master/arch/sh/mm/ |
H A D | kmap.c | 21 unsigned long vaddr; local 24 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); 25 kmap_coherent_pte = virt_to_kpte(vaddr); 32 unsigned long vaddr; local 43 vaddr = __fix_to_virt(idx); 48 return (void *)vaddr; 54 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; local 55 enum fixed_addresses idx = __virt_to_fix(vaddr); 58 __flush_purge_region((void *)vaddr, PAGE_SIZE); 60 pte_clear(&init_mm, vaddr, kmap_coherent_pt 61 local_flush_tlb_one(get_asid(), vaddr); local [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | tlbflush_64.h | 21 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr, 51 void __flush_tlb_page(unsigned long context, unsigned long vaddr); 56 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) argument 58 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); 64 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr); 66 #define global_flush_tlb_page(mm, vaddr) \ 67 smp_flush_tlb_page(mm, vaddr)
|
/linux-master/arch/csky/mm/ |
H A D | highmem.c | 22 unsigned long vaddr; local 28 vaddr = PKMAP_BASE; 29 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, swapper_pg_dir); 31 pgd = swapper_pg_dir + pgd_index(vaddr); 33 pmd = pmd_offset(pud, vaddr); 34 pte = pte_offset_kernel(pmd, vaddr);
|
/linux-master/drivers/net/ethernet/freescale/fman/ |
H A D | fman_muram.c | 20 unsigned long vaddr) 22 return vaddr - (unsigned long)muram->vbase; 41 void __iomem *vaddr; local 54 vaddr = ioremap(base, size); 55 if (!vaddr) { 60 ret = gen_pool_add_virt(muram->pool, (unsigned long)vaddr, 64 iounmap(vaddr); 68 memset_io(vaddr, 0, (int)size); 70 muram->vbase = vaddr; 107 unsigned long vaddr; local 19 fman_muram_vbase_to_offset(struct muram_info *muram, unsigned long vaddr) argument [all...] |
/linux-master/arch/riscv/mm/ |
H A D | kasan_init.c | 29 static void __init kasan_populate_pte(pmd_t *pmd, unsigned long vaddr, unsigned long end) argument 39 ptep = pte_offset_kernel(pmd, vaddr); 47 } while (ptep++, vaddr += PAGE_SIZE, vaddr != end); 50 static void __init kasan_populate_pmd(pud_t *pud, unsigned long vaddr, unsigned long end) argument 61 pmdp = pmd_offset(pud, vaddr); 64 next = pmd_addr_end(vaddr, end); 66 if (pmd_none(pmdp_get(pmdp)) && IS_ALIGNED(vaddr, PMD_SIZE) && 67 (next - vaddr) >= PMD_SIZE) { 76 kasan_populate_pte(pmdp, vaddr, nex 80 kasan_populate_pud(p4d_t *p4d, unsigned long vaddr, unsigned long end) argument 111 kasan_populate_p4d(pgd_t *pgd, unsigned long vaddr, unsigned long end) argument 142 kasan_populate_pgd(pgd_t *pgdp, unsigned long vaddr, unsigned long end) argument 165 kasan_early_clear_pud(p4d_t *p4dp, unsigned long vaddr, unsigned long end) argument 190 kasan_early_clear_p4d(pgd_t *pgdp, unsigned long vaddr, unsigned long end) argument 216 kasan_early_clear_pgd(pgd_t *pgdp, unsigned long vaddr, unsigned long end) argument 234 kasan_early_populate_pud(p4d_t *p4dp, unsigned long vaddr, unsigned long end) argument 263 kasan_early_populate_p4d(pgd_t *pgdp, unsigned long vaddr, unsigned long end) argument 301 kasan_early_populate_pgd(pgd_t *pgdp, unsigned long vaddr, unsigned long end) argument 371 unsigned long vaddr = (unsigned long)start & PAGE_MASK; local 374 kasan_populate_pgd(pgd_offset_k(vaddr), vaddr, vend); local 377 kasan_shallow_populate_pud(p4d_t *p4d, unsigned long vaddr, unsigned long end) argument 397 kasan_shallow_populate_p4d(pgd_t *pgd, unsigned long vaddr, unsigned long end) argument 417 kasan_shallow_populate_pgd(unsigned long vaddr, unsigned long end) argument 438 unsigned long vaddr = (unsigned long)start & PAGE_MASK; local [all...] |
/linux-master/arch/x86/xen/ |
H A D | mmu.c | 11 unsigned long arbitrary_virt_to_mfn(void *vaddr) argument 13 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); 18 xmaddr_t arbitrary_virt_to_machine(void *vaddr) argument 20 unsigned long address = (unsigned long)vaddr; 26 * if the PFN is in the linear mapped vaddr range, we can just use 29 if (virt_addr_valid(vaddr)) 30 return virt_to_machine(vaddr);
|
/linux-master/arch/arm/include/asm/ |
H A D | highmem.h | 66 #define arch_kmap_local_post_map(vaddr, pteval) \ 67 local_flush_tlb_kernel_page(vaddr) 69 #define arch_kmap_local_pre_unmap(vaddr) \ 72 __cpuc_flush_dcache_area((void *)vaddr, PAGE_SIZE); \ 75 #define arch_kmap_local_post_unmap(vaddr) \ 76 local_flush_tlb_kernel_page(vaddr)
|
/linux-master/arch/powerpc/include/asm/ |
H A D | highmem.h | 61 #define arch_kmap_local_set_pte(mm, vaddr, ptep, ptev) \ 62 __set_pte_at(mm, vaddr, ptep, ptev, 1) 63 #define arch_kmap_local_post_map(vaddr, pteval) \ 64 local_flush_tlb_page(NULL, vaddr) 65 #define arch_kmap_local_post_unmap(vaddr) \ 66 local_flush_tlb_page(NULL, vaddr)
|
/linux-master/arch/mips/mm/ |
H A D | pgtable-32.c | 53 unsigned long vaddr; local 72 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1); 73 fixrange_init(vaddr & PMD_MASK, vaddr + FIXADDR_SIZE, pgd_base); 79 vaddr = PKMAP_BASE; 80 fixrange_init(vaddr & PMD_MASK, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); 82 pgd = swapper_pg_dir + pgd_index(vaddr); 83 p4d = p4d_offset(pgd, vaddr); 84 pud = pud_offset(p4d, vaddr); [all...] |
/linux-master/arch/microblaze/include/asm/ |
H A D | highmem.h | 54 #define arch_kmap_local_post_map(vaddr, pteval) \ 55 local_flush_tlb_page(NULL, vaddr); 56 #define arch_kmap_local_post_unmap(vaddr) \ 57 local_flush_tlb_page(NULL, vaddr);
|
/linux-master/arch/csky/include/asm/ |
H A D | highmem.h | 37 #define arch_kmap_local_post_map(vaddr, pteval) kmap_flush_tlb(vaddr) 38 #define arch_kmap_local_post_unmap(vaddr) kmap_flush_tlb(vaddr)
|