Searched refs:PAGE_OFFSET (Results 151 - 175 of 205) sorted by relevance

123456789

/linux-master/arch/s390/include/asm/
H A Dpage.h209 #define PAGE_OFFSET __PAGE_OFFSET macro
/linux-master/arch/arm/kernel/
H A Dphys2virt.S23 * PHYS_OFFSET and PAGE_OFFSET, which is assumed to be
36 subs r3, r8, #PAGE_OFFSET @ PHYS_OFFSET - PAGE_OFFSET
/linux-master/arch/xtensa/mm/
H A Dmisc.S123 movi a5, PAGE_OFFSET
/linux-master/arch/powerpc/mm/book3s64/
H A Dslb.c578 create_shadowed_slbe(PAGE_OFFSET, mmu_kernel_ssize, lflags, LINEAR_INDEX);
588 (get_paca()->kstack & slb_esid_mask(mmu_kernel_ssize)) > PAGE_OFFSET)
/linux-master/arch/powerpc/kernel/
H A Dhead_64.S80 * after prom and kexec entry run at the virtual address (PAGE_OFFSET).
349 /* Now NIA and r2 are relocated to PAGE_OFFSET if not already */
387 /* Now the MMU is off, can branch to our PAGE_OFFSET address */
510 /* Now the MMU is off, can return to our PAGE_OFFSET address */
557 /* These functions return to the virtual (PAGE_OFFSET) address */
633 LOAD_REG_IMMEDIATE(r25, PAGE_OFFSET) /* else use static kernel base */
649 LOAD_REG_IMMEDIATE(r3, PAGE_OFFSET)
779 /* Branch to our PAGE_OFFSET address */
H A Dhead_85xx.S94 * from device tree. So we will map PAGE_OFFSET to memstart_addr,
96 * PAGE_OFFSET + (kernstart_addr - memstart_addr)
101 lis r3,PAGE_OFFSET@h
119 * We could map the 64M page we belong to at PAGE_OFFSET and
449 lis r11, PAGE_OFFSET@h
557 lis r11, PAGE_OFFSET@h
700 * CR5 - results of addr >= PAGE_OFFSET
1178 * r4 - __pa(PAGE_OFFSET in AS1) - __pa(PAGE_OFFSET in AS0)
1190 * We may map the PAGE_OFFSET i
[all...]
/linux-master/fs/proc/
H A Dkcore.c47 #define kc_vaddr_to_offset(v) ((v) - PAGE_OFFSET)
50 #define kc_offset_to_vaddr(o) ((o) + PAGE_OFFSET)
/linux-master/arch/x86/kernel/
H A Dsetup.c269 initrd_start = relocated_ramdisk + PAGE_OFFSET;
315 initrd_start = ramdisk_image + PAGE_OFFSET;
/linux-master/arch/sparc/kernel/
H A Detrap_32.S234 sethi %hi(PAGE_OFFSET), %glob_tmp
H A Dwof.S330 sethi %hi(PAGE_OFFSET), %glob_tmp
H A Drtrap_32.S231 sethi %hi(PAGE_OFFSET), %g1
H A Dtraps_32.c81 (((unsigned long) rw) >= PAGE_OFFSET) &&
/linux-master/arch/sparc/mm/
H A Dfault_32.c146 if (!from_user && address >= PAGE_OFFSET)
H A Dsrmmu.c864 const unsigned long min_vaddr = PAGE_OFFSET;
865 const unsigned long max_vaddr = PAGE_OFFSET + SRMMU_MAXMEM;
885 do_large_mapping(PAGE_OFFSET, phys_base);
/linux-master/arch/riscv/kernel/
H A Dhibernate.c357 unsigned long start = PAGE_OFFSET;
/linux-master/arch/sparc/include/asm/
H A Dpgtable_32.h50 #define USER_PTRS_PER_PGD PAGE_OFFSET / PGDIR_SIZE
/linux-master/arch/microblaze/include/asm/
H A Dpgtable.h104 #define USER_PGD_PTRS (PAGE_OFFSET >> PGDIR_SHIFT)
/linux-master/arch/arm64/mm/
H A Dmmu.c510 if (virt < PAGE_OFFSET) {
537 if (virt < PAGE_OFFSET) {
1298 WARN_ON((start < PAGE_OFFSET) || (end > PAGE_END));
1301 free_empty_tables(start, end, PAGE_OFFSET, PAGE_END);
1325 * Linear mapping region is the range [PAGE_OFFSET..(PAGE_END - 1)]
H A Dptdump.c363 { PAGE_OFFSET, "Linear Mapping start" },
/linux-master/arch/arm/mm/
H A Dmmu.c991 md->virtual >= PAGE_OFFSET && md->virtual < FIXADDR_START &&
1182 vmalloc_max = VMALLOC_END - (PAGE_OFFSET + SZ_32M + VMALLOC_OFFSET);
1210 PAGE_OFFSET + PHYS_OFFSET;
1323 for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE)
/linux-master/arch/arm64/include/asm/
H A Dprocessor.h334 return arch_kasan_reset_tag(addr) >= PAGE_OFFSET;
/linux-master/arch/x86/mm/
H A Dpti.c176 if (address < PAGE_OFFSET) {
/linux-master/arch/riscv/mm/
H A Dptdump.c394 address_markers[PAGE_OFFSET_NR].start_address = PAGE_OFFSET;
/linux-master/arch/x86/mm/pat/
H A Dset_memory.c1611 if (within(vaddr, PAGE_OFFSET,
1612 PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT))) {
1727 if (!(within(vaddr, PAGE_OFFSET,
1728 PAGE_OFFSET + (max_pfn_mapped << PAGE_SHIFT)))) {
2080 decoy_addr = (pfn << PAGE_SHIFT) + (PAGE_OFFSET ^ BIT(63));
/linux-master/arch/m68k/mm/
H A Dmotorola.c460 m68k_memoffset = min_addr - PAGE_OFFSET;

Completed in 341 milliseconds

123456789