/linux-master/include/asm-generic/ |
H A D | logic_io.h | 37 void iounmap(void volatile __iomem *addr); 40 u8 __raw_readb(const volatile void __iomem *addr); 43 u16 __raw_readw(const volatile void __iomem *addr); 46 u32 __raw_readl(const volatile void __iomem *addr); 50 u64 __raw_readq(const volatile void __iomem *addr); 54 void __raw_writeb(u8 value, volatile void __iomem *addr); 57 void __raw_writew(u16 value, volatile void __iomem *addr); 60 void __raw_writel(u32 value, volatile void __iomem *addr); 64 void __raw_writeq(u64 value, volatile void __iomem *addr); 68 void memset_io(volatile void __iomem *addr, in [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | hugetlb.h | 16 unsigned long addr, 24 if (addr & ~huge_page_mask(h)) 28 if (task_size - len < addr) 35 unsigned long addr, pte_t *ptep) 41 set_pte_at(mm, addr, ptep, clear); 47 unsigned long addr, pte_t *ptep) 56 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); 57 flush_tlb_page(vma, addr); 70 unsigned long addr, 77 set_pte_at(vma->vm_mm, addr, pte 15 prepare_hugepage_range(struct file *file, unsigned long addr, unsigned long len) argument 34 huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 46 huge_ptep_clear_flush(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 69 huge_ptep_set_access_flags(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t pte, int dirty) argument [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | hugetlb.h | 15 unsigned long addr, 23 if (addr & ~huge_page_mask(h)) 27 if (task_size - len < addr) 34 unsigned long addr, pte_t *ptep) 40 set_pte_at(mm, addr, ptep, clear); 46 unsigned long addr, pte_t *ptep) 50 pte = huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); 51 flush_tlb_page(vma, addr); 64 unsigned long addr, 71 set_pte_at(vma->vm_mm, addr, pte 14 prepare_hugepage_range(struct file *file, unsigned long addr, unsigned long len) argument 33 huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 45 huge_ptep_clear_flush(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument 63 huge_ptep_set_access_flags(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep, pte_t pte, int dirty) argument [all...] |
/linux-master/arch/sh/mm/ |
H A D | hugetlbpage.c | 25 unsigned long addr, unsigned long sz) 33 pgd = pgd_offset(mm, addr); 35 p4d = p4d_alloc(mm, pgd, addr); 37 pud = pud_alloc(mm, p4d, addr); 39 pmd = pmd_alloc(mm, pud, addr); 41 pte = pte_alloc_huge(mm, pmd, addr); 50 unsigned long addr, unsigned long sz) 58 pgd = pgd_offset(mm, addr); 60 p4d = p4d_offset(pgd, addr); 62 pud = pud_offset(p4d, addr); 24 huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long sz) argument 49 huge_pte_offset(struct mm_struct *mm, unsigned long addr, unsigned long sz) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | ioremap_32.c | 9 void __iomem *ioremap_wt(phys_addr_t addr, unsigned long size) argument 13 return __ioremap_caller(addr, size, prot, __builtin_return_address(0)); 18 __ioremap_caller(phys_addr_t addr, unsigned long size, pgprot_t prot, void *caller) argument 28 if (addr < SZ_16M) 29 addr += _ISA_MEM_BASE; 37 p = addr & PAGE_MASK; 38 offset = addr & ~PAGE_MASK; 39 size = PAGE_ALIGN(addr + size) - p; 48 pr_warn("%s(): phys addr 0x%llx is RAM lr %ps\n", __func__, 66 return generic_ioremap_prot(addr, siz 81 iounmap(volatile void __iomem *addr) argument [all...] |
/linux-master/include/linux/ |
H A D | ioremap.h | 19 unsigned long addr = (unsigned long)kasan_reset_tag(x); local 21 return addr >= IOREMAP_START && addr < IOREMAP_END;
|
H A D | kfence.h | 35 * @addr: address to check 51 static __always_inline bool is_kfence_address(const void *addr) argument 55 * where __kfence_pool == NULL && addr < KFENCE_POOL_SIZE. Keep it in 58 return unlikely((unsigned long)((char *)addr - __kfence_pool) < KFENCE_POOL_SIZE && __kfence_pool); 134 * @addr: pointer to a heap object 144 size_t kfence_ksize(const void *addr); 148 * @addr: address within a KFENCE-allocated object 158 void *kfence_object_start(const void *addr); 162 * @addr: object to be freed 164 * Requires: is_kfence_address(addr) 183 kfence_free(void *addr) argument 228 is_kfence_address(const void *addr) argument 233 kfence_ksize(const void *addr) argument 234 kfence_object_start(const void *addr) argument 235 __kfence_free(void *addr) argument 236 kfence_free(void *addr) argument 237 kfence_handle_page_fault(unsigned long addr, bool is_write, struct pt_regs *regs) argument [all...] |
H A D | extable.h | 23 search_kernel_exception_table(unsigned long addr); 27 const struct exception_table_entry *search_module_extables(unsigned long addr); 30 search_module_extables(unsigned long addr) argument 37 const struct exception_table_entry *search_bpf_extables(unsigned long addr); 40 search_bpf_extables(unsigned long addr) argument
|
/linux-master/arch/s390/include/asm/ |
H A D | dma-types.h | 43 static inline void *dma32_to_virt(dma32_t addr) argument 45 return __va((__force unsigned long)addr); 48 static inline dma32_t u32_to_dma32(u32 addr) argument 50 return (__force dma32_t)addr; 53 static inline u32 dma32_to_u32(dma32_t addr) argument 55 return (__force u32)addr; 78 static inline void *dma64_to_virt(dma64_t addr) argument 80 return __va((__force unsigned long)addr); 83 static inline dma64_t u64_to_dma64(u64 addr) argument 85 return (__force dma64_t)addr; 88 dma64_to_u64(dma64_t addr) argument [all...] |
/linux-master/mm/kasan/ |
H A D | init.c | 92 static void __ref zero_pte_populate(pmd_t *pmd, unsigned long addr, argument 95 pte_t *pte = pte_offset_kernel(pmd, addr); 102 while (addr + PAGE_SIZE <= end) { 103 set_pte_at(&init_mm, addr, pte, zero_pte); 104 addr += PAGE_SIZE; 105 pte = pte_offset_kernel(pmd, addr); 109 static int __ref zero_pmd_populate(pud_t *pud, unsigned long addr, argument 112 pmd_t *pmd = pmd_offset(pud, addr); 116 next = pmd_addr_end(addr, end); 118 if (IS_ALIGNED(addr, PMD_SIZ 142 pmd_init(void *addr) argument 146 zero_pud_populate(p4d_t *p4d, unsigned long addr, unsigned long end) argument 184 pud_init(void *addr) argument 188 zero_p4d_populate(pgd_t *pgd, unsigned long addr, unsigned long end) argument 239 unsigned long addr = (unsigned long)shadow_start; local 352 kasan_remove_pte_table(pte_t *pte, unsigned long addr, unsigned long end) argument 374 kasan_remove_pmd_table(pmd_t *pmd, unsigned long addr, unsigned long end) argument 400 kasan_remove_pud_table(pud_t *pud, unsigned long addr, unsigned long end) argument 427 kasan_remove_p4d_table(p4d_t *p4d, unsigned long addr, unsigned long end) argument 455 unsigned long addr, end, next; local [all...] |
/linux-master/arch/powerpc/include/asm/ |
H A D | io.h | 98 static inline u##size name(const volatile u##size __iomem *addr) \ 102 : "=r" (ret) : "r" (addr) : "memory"); \ 107 static inline void name(volatile u##size __iomem *addr, u##size val) \ 110 : : "r" (addr), "r" (val) : "memory"); \ 115 static inline u##size name(const volatile u##size __iomem *addr) \ 119 : "=r" (ret) : "b" (addr) : "memory"); \ 124 static inline void name(volatile u##size __iomem *addr, u##size val) \ 127 : : "b" (addr), "r" (val) : "memory"); \ 132 static inline u##size name(const volatile u##size __iomem *addr) \ 136 : "=r" (ret) : "Z" (*addr) 199 in_le64(const volatile u64 __iomem *addr) argument 204 out_le64(volatile u64 __iomem *addr, u64 val) argument 213 in_be64(const volatile u64 __iomem *addr) argument 218 out_be64(volatile u64 __iomem *addr, u64 val) argument 330 __raw_readb(const volatile void __iomem *addr) argument 336 __raw_readw(const volatile void __iomem *addr) argument 342 __raw_readl(const volatile void __iomem *addr) argument 348 __raw_writeb(unsigned char v, volatile void __iomem *addr) argument 354 __raw_writew(unsigned short v, volatile void __iomem *addr) argument 360 __raw_writel(unsigned int v, volatile void __iomem *addr) argument 367 __raw_readq(const volatile void __iomem *addr) argument 373 __raw_writeq(unsigned long v, volatile void __iomem *addr) argument 379 __raw_writeq_be(unsigned long v, volatile void __iomem *addr) argument 728 ioread16be(const void __iomem *addr) argument 734 ioread32be(const void __iomem *addr) argument 741 ioread64_lo_hi(const void __iomem *addr) argument 747 ioread64_hi_lo(const void __iomem *addr) argument 753 ioread64be(const void __iomem *addr) argument 759 ioread64be_lo_hi(const void __iomem *addr) argument 765 ioread64be_hi_lo(const void __iomem *addr) argument 772 iowrite16be(u16 val, void __iomem *addr) argument 778 iowrite32be(u32 val, void __iomem *addr) argument 785 iowrite64_lo_hi(u64 val, void __iomem *addr) argument 791 iowrite64_hi_lo(u64 val, void __iomem *addr) argument 797 iowrite64be(u64 val, void __iomem *addr) argument 803 iowrite64be_lo_hi(u64 val, void __iomem *addr) argument 809 iowrite64be_hi_lo(u64 val, void __iomem *addr) argument [all...] |
/linux-master/arch/arm/kernel/ |
H A D | patch.c | 15 void *addr; member in struct:patch 22 static void __kprobes *patch_map(void *addr, int fixmap, unsigned long *flags) argument 24 unsigned int uintaddr = (uintptr_t) addr; 29 page = vmalloc_to_page(addr); 31 page = virt_to_page(addr); 33 return addr; 51 static void __kprobes *patch_map(void *addr, int fixmap, unsigned long *flags) argument 53 return addr; 58 void __kprobes __patch_text_real(void *addr, unsigned int insn, bool remap) argument 61 unsigned int uintaddr = (uintptr_t) addr; 120 patch_text(void *addr, unsigned int insn) argument [all...] |
/linux-master/arch/loongarch/mm/ |
H A D | mmap.c | 14 #define COLOUR_ALIGN(addr, pgoff) \ 15 ((((addr) + SHM_ALIGN_MASK) & ~SHM_ALIGN_MASK) \ 26 unsigned long addr = addr0; local 35 if (TASK_SIZE - len < addr) 43 ((addr - (pgoff << PAGE_SHIFT)) & SHM_ALIGN_MASK)) 45 return addr; 53 if (addr) { 55 addr = COLOUR_ALIGN(addr, pgoff); 57 addr 129 valid_phys_addr_range(phys_addr_t addr, size_t size) argument [all...] |
/linux-master/arch/riscv/mm/ |
H A D | fault.c | 25 static void die_kernel_fault(const char *msg, unsigned long addr, argument 31 addr); 38 static inline void no_context(struct pt_regs *regs, unsigned long addr) argument 50 if (addr < PAGE_SIZE) 53 if (kfence_handle_page_fault(addr, regs->cause == EXC_STORE_PAGE_FAULT, regs)) 59 die_kernel_fault(msg, addr, regs); 62 static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault) argument 70 no_context(regs, addr); 78 no_context(regs, addr); 81 do_trap(regs, SIGBUS, BUS_ADRERR, addr); 88 bad_area_nosemaphore(struct pt_regs *regs, int code, unsigned long addr) argument 104 bad_area(struct pt_regs *regs, struct mm_struct *mm, int code, unsigned long addr) argument 112 vmalloc_fault(struct pt_regs *regs, int code, unsigned long addr) argument 227 unsigned long addr, cause; local [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | hugetlb-madvise.c | 40 void write_fault_pages(void *addr, unsigned long nr_pages) argument 45 *((unsigned long *)(addr + (i * huge_page_size))) = i; 48 void read_fault_pages(void *addr, unsigned long nr_pages) argument 54 dummy += *((unsigned long *)(addr + (i * huge_page_size))); 64 void *addr, *addr2; local 92 * Test validity of MADV_DONTNEED addr and length arguments. mmap 97 addr = mmap(NULL, (NR_HUGE_PAGES + 2) * huge_page_size, 101 if (addr == MAP_FAILED) { 105 if (munmap(addr, huge_page_size) || 106 munmap(addr [all...] |
H A D | map_hugetlb.c | 33 static void check_bytes(char *addr) argument 35 ksft_print_msg("First hex is %x\n", *((unsigned int *)addr)); 38 static void write_bytes(char *addr, size_t length) argument 43 *(addr + i) = (char)i; 46 static void read_bytes(char *addr, size_t length) argument 50 check_bytes(addr); 52 if (*(addr + i) != (char)i) 60 void *addr; local 88 addr = mmap(ADDR, length, PROTECTION, flags, -1, 0); 89 if (addr [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | io.h | 62 static inline void __ ## port ## len(const volatile void __iomem *addr, \ 71 ctype x = __raw_read ## len(addr); \ 79 static inline void __ ## port ## len(volatile void __iomem *addr, \ 88 __raw_write ## len(*buf++, addr); \ 94 __io_reads_ins(reads, u8, b, __io_br(), __io_ar(addr)) 95 __io_reads_ins(reads, u16, w, __io_br(), __io_ar(addr)) 96 __io_reads_ins(reads, u32, l, __io_br(), __io_ar(addr)) 97 #define readsb(addr, buffer, count) __readsb(addr, buffer, count) 98 #define readsw(addr, buffe [all...] |
/linux-master/arch/sh/kernel/ |
H A D | sys_sh.c | 32 asmlinkage int old_mmap(unsigned long addr, unsigned long len, argument 38 return ksys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT); 41 asmlinkage long sys_mmap2(unsigned long addr, unsigned long len, argument 54 return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); 58 asmlinkage int sys_cacheflush(unsigned long addr, unsigned long len, int op) argument 69 if (addr + len < addr) 73 vma = find_vma (current->mm, addr); 74 if (vma == NULL || addr < vma->vm_start || addr [all...] |
/linux-master/arch/sparc/include/asm/ |
H A D | bitops_32.h | 22 unsigned long sp32___set_bit(unsigned long *addr, unsigned long mask); 23 unsigned long sp32___clear_bit(unsigned long *addr, unsigned long mask); 24 unsigned long sp32___change_bit(unsigned long *addr, unsigned long mask); 27 * Set bit 'nr' in 32-bit quantity at address 'addr' where bit '0' 32 static inline int test_and_set_bit(unsigned long nr, volatile unsigned long *addr) argument 36 ADDR = ((unsigned long *) addr) + (nr >> 5); 42 static inline void set_bit(unsigned long nr, volatile unsigned long *addr) argument 46 ADDR = ((unsigned long *) addr) + (nr >> 5); 52 static inline int test_and_clear_bit(unsigned long nr, volatile unsigned long *addr) argument 56 ADDR = ((unsigned long *) addr) 62 clear_bit(unsigned long nr, volatile unsigned long *addr) argument 72 test_and_change_bit(unsigned long nr, volatile unsigned long *addr) argument 82 change_bit(unsigned long nr, volatile unsigned long *addr) argument [all...] |
/linux-master/arch/sh/include/asm/ |
H A D | bitops-op32.h | 22 arch___set_bit(unsigned long nr, volatile unsigned long *addr) argument 27 : "+r" (addr) 33 unsigned long *p = ((unsigned long *)addr) + BIT_WORD(nr); 40 arch___clear_bit(unsigned long nr, volatile unsigned long *addr) argument 45 : "+r" (addr) 52 unsigned long *p = ((unsigned long *)addr) + BIT_WORD(nr); 61 * @addr: the address to start counting from 68 arch___change_bit(unsigned long nr, volatile unsigned long *addr) argument 73 : "+r" (addr) 80 unsigned long *p = ((unsigned long *)addr) 96 arch___test_and_set_bit(unsigned long nr, volatile unsigned long *addr) argument 116 arch___test_and_clear_bit(unsigned long nr, volatile unsigned long *addr) argument 128 arch___test_and_change_bit(unsigned long nr, volatile unsigned long *addr) argument [all...] |
/linux-master/arch/x86/mm/ |
H A D | ident_map.c | 8 unsigned long addr, unsigned long end) 10 addr &= PMD_MASK; 11 for (; addr < end; addr += PMD_SIZE) { 12 pmd_t *pmd = pmd_page + pmd_index(addr); 17 set_pmd(pmd, __pmd((addr - info->offset) | info->page_flag)); 22 unsigned long addr, unsigned long end) 26 for (; addr < end; addr = next) { 27 pud_t *pud = pud_page + pud_index(addr); 7 ident_pmd_init(struct x86_mapping_info *info, pmd_t *pmd_page, unsigned long addr, unsigned long end) argument 21 ident_pud_init(struct x86_mapping_info *info, pud_t *pud_page, unsigned long addr, unsigned long end) argument 61 ident_p4d_init(struct x86_mapping_info *info, p4d_t *p4d_page, unsigned long addr, unsigned long end) argument 100 unsigned long addr = pstart + info->offset; local [all...] |
/linux-master/arch/csky/kernel/ |
H A D | syscall.c | 6 SYSCALL_DEFINE1(set_thread_area, unsigned long, addr) 11 reg->tls = addr; 12 ti->tp_value = addr; 18 unsigned long, addr, 28 return ksys_mmap_pgoff(addr, len, prot, flags, fd,
|
/linux-master/drivers/s390/cio/ |
H A D | ioasm.h | 16 int stsch(struct subchannel_id schid, struct schib *addr); 17 int msch(struct subchannel_id schid, struct schib *addr); 18 int tsch(struct subchannel_id schid, struct irb *addr); 19 int ssch(struct subchannel_id schid, union orb *addr); 21 int tpi(struct tpi_info *addr);
|
/linux-master/drivers/staging/rtl8712/ |
H A D | rtl8712_io.c | 25 u8 r8712_read8(struct _adapter *adapter, u32 addr) argument 29 return hdl->io_ops._read8(hdl, addr); 32 u16 r8712_read16(struct _adapter *adapter, u32 addr) argument 36 return hdl->io_ops._read16(hdl, addr); 39 u32 r8712_read32(struct _adapter *adapter, u32 addr) argument 43 return hdl->io_ops._read32(hdl, addr); 46 void r8712_write8(struct _adapter *adapter, u32 addr, u8 val) argument 50 hdl->io_ops._write8(hdl, addr, val); 53 void r8712_write16(struct _adapter *adapter, u32 addr, u16 val) argument 57 hdl->io_ops._write16(hdl, addr, va 60 r8712_write32(struct _adapter *adapter, u32 addr, u32 val) argument 67 r8712_read_mem(struct _adapter *adapter, u32 addr, u32 cnt, u8 *pmem) argument 77 r8712_write_mem(struct _adapter *adapter, u32 addr, u32 cnt, u8 *pmem) argument 84 r8712_read_port(struct _adapter *adapter, u32 addr, u32 cnt, u8 *pmem) argument 94 r8712_write_port(struct _adapter *adapter, u32 addr, u32 cnt, u8 *pmem) argument [all...] |
/linux-master/arch/arm/mach-alpine/ |
H A D | platsmp.c | 19 phys_addr_t addr; local 21 addr = __pa_symbol(secondary_startup); 23 if (addr > (phys_addr_t)(uint32_t)(-1)) { 24 pr_err("FAIL: resume address over 32bit (%pa)", &addr); 28 return alpine_cpu_wakeup(cpu_logical_map(cpu), (uint32_t)addr);
|