/linux-master/arch/csky/include/asm/ |
H A D | tlbflush.h | 12 * - flush_tlb_range(vma, start, end) flushes a range of pages 13 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 19 unsigned long end); 20 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
/linux-master/arch/csky/kernel/ |
H A D | vdso.c | 58 goto end; 75 goto end; 84 end:
|
/linux-master/arch/csky/mm/ |
H A D | cachev1.c | 45 unsigned int end, 53 if (unlikely((end - start) >= PAGE_SIZE) || 68 for (; i < end; i += L1_CACHE_BYTES) { 87 void icache_inv_range(unsigned long start, unsigned long end) argument 89 cache_op_range(start, end, INS_CACHE|CACHE_INV, 0); 102 void dcache_wb_range(unsigned long start, unsigned long end) argument 104 cache_op_range(start, end, DATA_CACHE|CACHE_CLR, 0); 112 void cache_wbinv_range(unsigned long start, unsigned long end) argument 114 cache_op_range(start, end, INS_CACHE|DATA_CACHE|CACHE_CLR|CACHE_INV, 0); 123 void dma_wbinv_range(unsigned long start, unsigned long end) argument 43 cache_op_range( unsigned int start, unsigned int end, unsigned int value, unsigned int l2) argument 128 dma_inv_range(unsigned long start, unsigned long end) argument 133 dma_wb_range(unsigned long start, unsigned long end) argument [all...] |
H A D | cachev2.c | 24 void icache_inv_range(unsigned long start, unsigned long end) argument 28 for (; i < end; i += L1_CACHE_BYTES) 35 unsigned long end; member in struct:cache_range 54 for (; i < param->end; i += L1_CACHE_BYTES) 62 void icache_inv_range(unsigned long start, unsigned long end) argument 64 struct cache_range param = { start, end }; 79 void dcache_wb_range(unsigned long start, unsigned long end) argument 83 for (; i < end; i += L1_CACHE_BYTES) 88 void cache_wbinv_range(unsigned long start, unsigned long end) argument 90 dcache_wb_range(start, end); 95 dma_wbinv_range(unsigned long start, unsigned long end) argument 104 dma_inv_range(unsigned long start, unsigned long end) argument 113 dma_wb_range(unsigned long start, unsigned long end) argument [all...] |
H A D | dma-mapping.c | 15 void (*fn)(unsigned long start, unsigned long end)) 47 static void dma_wbinv_set_zero_range(unsigned long start, unsigned long end) argument 49 memset((void *)start, 0, end - start); 50 dma_wbinv_range(start, end); 14 cache_op(phys_addr_t paddr, size_t size, void (*fn)(unsigned long start, unsigned long end)) argument
|
H A D | init.c | 56 pr_err("initrd extends beyond end of memory"); 154 void __init fixrange_init(unsigned long start, unsigned long end, argument 170 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { 172 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { 174 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) {
|
H A D | tlb.c | 53 unsigned long end) 58 end += TLB_ENTRY_SIZE - 1; 59 end &= TLB_ENTRY_SIZE_MASK; 63 while (start < end) { 79 while (start < end) { 95 void flush_tlb_kernel_range(unsigned long start, unsigned long end) argument 98 end += TLB_ENTRY_SIZE - 1; 99 end &= TLB_ENTRY_SIZE_MASK; 103 while (start < end) { 119 while (start < end) { 52 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument [all...] |
/linux-master/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 18 * - flush_cache_range(vma, start, end) flushes a range of pages 19 * - flush_icache_range(start, end) flush a range of instructions 31 extern void flush_dcache_range(unsigned long start, unsigned long end); 37 extern void flush_icache_range(unsigned long start, unsigned long end); 78 extern void hexagon_inv_dcache_range(unsigned long start, unsigned long end); 79 extern void hexagon_clean_dcache_range(unsigned long start, unsigned long end);
|
H A D | io.h | 28 unsigned long end, unsigned long flags);
|
/linux-master/arch/hexagon/kernel/ |
H A D | time.c | 49 .end = RTOS_TIMER_REGS_ADDR+PAGE_SIZE-1,
|
/linux-master/arch/hexagon/lib/ |
H A D | checksum.c | 88 int i, start, mid, end, mask; local 101 end = mid & 0xF; 148 ptr4 = (unsigned int *) &ptr[start + (mid * 16) + (end & 8)]; 149 if (end & 4) { 155 ptr2 = (unsigned short *) &ptr[start + (mid * 16) + (end & 12)]; 156 if (end & 2) 159 if (end & 1) 160 sum1 += (u64) ptr[start + (mid * 16) + (end & 14)]; 163 if (end & 8) {
|
/linux-master/arch/hexagon/mm/ |
H A D | vm_tlb.c | 27 unsigned long end) 32 __vmclrmap((void *)start, end - start); 50 /* should probably use that fixaddr end or whateve label */ 79 void flush_tlb_kernel_range(unsigned long start, unsigned long end) argument 81 __vmclrmap((void *)start, end - start); 26 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
|
/linux-master/arch/loongarch/include/asm/ |
H A D | alternative.h | 28 extern void apply_alternatives(struct alt_instr *start, struct alt_instr *end);
|
H A D | cacheflush.h | 35 void local_flush_icache_range(unsigned long start, unsigned long end); 43 #define flush_cache_range(vma, start, end) do { } while (0) 45 #define flush_cache_vmap(start, end) do { } while (0) 46 #define flush_cache_vunmap(start, end) do { } while (0)
|
H A D | kvm_host.h | 239 int kvm_unmap_hva_range(struct kvm *kvm, unsigned long start, unsigned long end, bool blockable); 240 int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end);
|
H A D | kvm_mmu.h | 103 phys_addr_t addr, phys_addr_t end) 109 return (boundary - 1 < end - 1) ? boundary : end; 102 kvm_pgtable_addr_end(kvm_ptw_ctx *ctx, phys_addr_t addr, phys_addr_t end) argument
|
H A D | numa.h | 27 u64 end; member in struct:numa_memblk 37 extern int __init numa_add_memblk(int nodeid, u64 start, u64 end);
|
H A D | stacktrace.h | 21 unsigned long begin, end, next_sp; member in struct:stack_info
|
H A D | tlb.h | 151 flush_tlb_range(&vma, tlb->start, tlb->end);
|
H A D | tlbflush.h | 16 * - flush_tlb_range(vma, start, end) flushes a range of pages 17 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 23 extern void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); 24 extern void local_flush_tlb_kernel_range(unsigned long start, unsigned long end); 41 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 42 #define flush_tlb_kernel_range(vmaddr, end) local_flush_tlb_kernel_range(vmaddr, end)
|
/linux-master/arch/loongarch/kernel/ |
H A D | acpi.c | 92 acpi_parse_processor(union acpi_subtable_headers *header, const unsigned long end) argument 97 if (BAD_MADT_ENTRY(processor, end)) 110 acpi_parse_eio_master(union acpi_subtable_headers *header, const unsigned long end) argument 116 if (BAD_MADT_ENTRY(eiointc, end))
|
H A D | alternative.c | 56 static inline bool in_alt_jump(unsigned long jump, void *start, void *end) argument 58 return jump >= (unsigned long)start && jump < (unsigned long)end; 63 void *start, void *end) 78 if (in_alt_jump(jump_addr, start, end)) 97 if (in_alt_jump(jump_addr, start, end)) 116 if (in_alt_jump(jump_addr, start, end)) 183 void __init_or_module apply_alternatives(struct alt_instr *start, struct alt_instr *end) argument 190 DPRINTK("alt table %px, -> %px", start, end); 192 * The scan order should be from start to end. A later scanned 200 for (a = start; a < end; 61 recompute_jump(union loongarch_instruction *buf, union loongarch_instruction *dest, union loongarch_instruction *src, void *start, void *end) argument [all...] |
H A D | dma.c | 11 u64 mask, end; local 16 end = dma_range_map_max(map); 18 mask = DMA_BIT_MASK(ilog2(end) + 1); 19 dev->bus_dma_limit = end;
|
H A D | numa.c | 151 static int __init numa_add_memblk_to(int nid, u64 start, u64 end, argument 155 if (start == end) 159 if (start > end || nid < 0 || nid >= MAX_NUMNODES) { 161 nid, start, end - 1); 171 mi->blk[mi->nr_blks].end = PFN_ALIGN(end - PAGE_SIZE + 1); 181 * @end: End address of the new memblk 188 int __init numa_add_memblk(int nid, u64 start, u64 end) argument 190 return numa_add_memblk_to(nid, start, end, &numa_meminfo); 250 static void __init add_numamem_region(u64 start, u64 end, u3 argument 340 phys_addr_t end = memblock_end_of_DRAM() - 1; local [all...] |
H A D | process.c | 236 unsigned long end = begin + IRQ_STACK_START; local 238 if (stack < begin || stack >= end) 241 nextsp = *(unsigned long *)end; 246 info->end = end; 257 unsigned long end = begin + THREAD_SIZE; local 259 if (stack < begin || stack >= end) 263 info->end = end;
|