/linux-master/tools/perf/arch/arm/tests/ |
H A D | vectors-page.c | 14 void *start, *end; local 16 if (find_map(&start, &end, VECTORS__MAP_NAME)) {
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | lru_bug.c | 16 goto end; 19 end:
|
/linux-master/drivers/acpi/arm64/ |
H A D | dma.c | 10 u64 end, mask; local 34 for (end = 0; r->size; r++) { 35 if (r->dma_start + r->size - 1 > end) 36 end = r->dma_start + r->size - 1; 39 size = end + 1; 50 end = size - 1; 51 mask = DMA_BIT_MASK(ilog2(end) + 1); 52 dev->bus_dma_limit = end;
|
/linux-master/arch/powerpc/include/asm/book3s/32/ |
H A D | tlbflush.h | 13 void hash__flush_range(struct mm_struct *mm, unsigned long start, unsigned long end); 26 * Called at the end of a mmu_gather operation to make sure the 36 static inline void flush_range(struct mm_struct *mm, unsigned long start, unsigned long end) argument 40 hash__flush_range(mm, start, end); 41 else if (end - start <= PAGE_SIZE) 64 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 66 flush_range(vma->vm_mm, start, end); 69 static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end) argument 71 flush_range(&init_mm, start, end);
|
/linux-master/arch/csky/abiv1/inc/abi/ |
H A D | cacheflush.h | 41 * if (current_mm != vma->mm) cache_wbinv_range(start, end) will be broken. 44 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); 45 #define flush_cache_vmap(start, end) cache_wbinv_all() 46 #define flush_cache_vmap_early(start, end) do { } while (0) 47 #define flush_cache_vunmap(start, end) cache_wbinv_all() 49 #define flush_icache_range(start, end) cache_wbinv_range(start, end) 50 #define flush_icache_mm_range(mm, start, end) cache_wbinv_range(start, end)
|
/linux-master/arch/sparc/include/asm/ |
H A D | tlbflush_64.h | 19 void flush_tsb_kernel_range(unsigned long start, unsigned long end); 36 unsigned long start, unsigned long end) 40 void flush_tlb_kernel_range(unsigned long start, unsigned long end); 52 void __flush_tlb_kernel_range(unsigned long start, unsigned long end); 63 void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end); 35 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
|
/linux-master/arch/x86/mm/pat/ |
H A D | memtype_interval.c | 43 return entry->end - 1; 57 static struct memtype *memtype_match(u64 start, u64 end, int match_type) argument 61 entry_match = interval_iter_first(&memtype_rbroot, start, end-1); 63 while (entry_match != NULL && entry_match->start < end) { 65 (entry_match->start == start) && (entry_match->end == end)) 69 (entry_match->start < start) && (entry_match->end == end)) 72 entry_match = interval_iter_next(entry_match, start, end-1); 78 static int memtype_check_conflict(u64 start, u64 end, argument 131 memtype_erase(u64 start, u64 end) argument [all...] |
/linux-master/arch/parisc/include/asm/ |
H A D | tlbflush.h | 17 unsigned long start, unsigned long end); 19 #define flush_tlb_range(vma, start, end) \ 20 __flush_tlb_range((vma)->vm_mm->context.space_id, start, end) 22 #define flush_tlb_kernel_range(start, end) \ 23 __flush_tlb_range(0, start, end)
|
/linux-master/net/sunrpc/auth_gss/ |
H A D | auth_gss_internal.h | 16 simple_get_bytes(const void *p, const void *end, void *res, size_t len) argument 19 if (unlikely(q > end || q < p)) 26 simple_get_netobj(const void *p, const void *end, struct xdr_netobj *dest) argument 31 p = simple_get_bytes(p, end, &len, sizeof(len)); 35 if (unlikely(q > end || q < p))
|
/linux-master/crypto/asymmetric_keys/ |
H A D | x509_loader.c | 12 const u8 *p, *end; local 16 end = p + list_size; 17 while (p < end) { 21 if (end - p < 4) 28 if (plen > end - p)
|
/linux-master/drivers/gpu/drm/arm/display/include/ |
H A D | malidp_utils.h | 26 /* the restriction of range is [start, end] */ 29 u32 end; member in struct:malidp_range 32 static inline void set_range(struct malidp_range *rg, u32 start, u32 end) argument 35 rg->end = end; 40 return (v >= rg->start) && (v <= rg->end);
|
/linux-master/arch/riscv/include/asm/ |
H A D | cacheflush.h | 35 #define flush_icache_range(start, end) flush_icache_all() 40 #define flush_cache_vmap(start, end) flush_tlb_kernel_range(start, end) 41 #define flush_cache_vmap_early(start, end) local_flush_tlb_kernel_range(start, end)
|
/linux-master/lib/ |
H A D | bitmap-str.c | 210 * start off group_len end nbits 216 unsigned int end; member in struct:region 224 for (start = r->start; start <= r->end; start += r->group_len) 225 bitmap_set(bitmap, start, min(r->end - start + 1, r->off)); 230 if (r->start > r->end || r->group_len == 0 || r->off > r->group_len) 233 if (r->end >= r->nbits) 287 static const char *bitmap_find_region_reverse(const char *start, const char *end) argument 289 while (start <= end && __end_of_region(*end)) 290 end 433 bitmap_get_x32_reverse(const char *start, const char *end, u32 *num) argument 476 const char *end = strnchrnul(start, buflen, '\\n') - 1; local [all...] |
/linux-master/arch/hexagon/mm/ |
H A D | cache.c | 12 #define spanlines(start, end) \ 13 (((end - (start & ~(LINESIZE - 1))) >> LINEBITS) + 1) 15 void flush_dcache_range(unsigned long start, unsigned long end) argument 17 unsigned long lines = spanlines(start, end-1); 35 void flush_icache_range(unsigned long start, unsigned long end) argument 37 unsigned long lines = spanlines(start, end-1); 60 void hexagon_clean_dcache_range(unsigned long start, unsigned long end) argument 62 unsigned long lines = spanlines(start, end-1); 80 void hexagon_inv_dcache_range(unsigned long start, unsigned long end) argument 82 unsigned long lines = spanlines(start, end [all...] |
/linux-master/arch/m68k/include/asm/ |
H A D | cacheflush_mm.h | 31 static inline void clear_cf_icache(unsigned long start, unsigned long end) argument 40 static inline void clear_cf_dcache(unsigned long start, unsigned long end) argument 49 static inline void clear_cf_bcache(unsigned long start, unsigned long end) argument 60 * The start and end addresses are cache line numbers not memory addresses. 62 static inline void flush_cf_icache(unsigned long start, unsigned long end) argument 66 for (set = start; set <= end; set += (0x10 - 3)) { 80 static inline void flush_cf_dcache(unsigned long start, unsigned long end) argument 84 for (set = start; set <= end; set += (0x10 - 3)) { 98 static inline void flush_cf_bcache(unsigned long start, unsigned long end) argument 102 for (set = start; set <= end; se 207 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 227 unsigned long addr, start, end; local [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | physmem_info.h | 18 u64 end; member in struct:physmem_range 36 unsigned long end; member in struct:reserved_range 61 void add_physmem_online_range(u64 start, u64 end); 64 unsigned long *end, bool respect_usable_limit) 68 *end = 0; 74 *end = (unsigned long)physmem_info.online[n].end; 77 *end = (unsigned long)physmem_info.online_extended[n - MEM_INLINED_ENTRIES].end; 83 if (*end > physmem_inf 63 __get_physmem_range(u32 n, unsigned long *start, unsigned long *end, bool respect_usable_limit) argument [all...] |
/linux-master/tools/workqueue/ |
H A D | wq_dump.py | 126 print(' pod_cpus', end='') 128 print(f' [{pod}]={cpumask_str(pt.pod_cpus[pod])}', end='') 131 print(' pod_node', end='') 133 print(f' [{pod}]={pt.pod_node[pod].value_()}', end='') 136 print(f' cpu_pod ', end='') 138 print(f' [{cpu}]={pt.cpu_pod[cpu].value_()}', end='') 159 print(f'pool[{pi:0{max_pool_id_len}}] flags=0x{pool.flags.value_():02x} ref={pool.refcnt.value_():{max_ref_len}} nice={pool.attrs.nice.value_():3} ', end='') 160 print(f'idle/workers={pool.nr_idle.value_():3}/{pool.nr_workers.value_():3} ', end='') 162 print(f'cpu={pool.cpu.value_():3}', end='') 164 print(' bh', end [all...] |
/linux-master/arch/sh/mm/ |
H A D | cache-sh2a.c | 53 unsigned long begin, end; local 58 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) 66 if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) { 68 end = begin + (nr_ways * current_cpu_data.dcache.way_size); 70 for (v = begin; v < end; v += L1_CACHE_BYTES) { 78 for (v = begin; v < end; v += L1_CACHE_BYTES) 94 unsigned long begin, end; local 98 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) 104 for (v = begin; v < end; v+=L1_CACHE_BYTES) { 124 unsigned long begin, end; local 153 unsigned long start, end; local [all...] |
H A D | cache-sh2.c | 21 unsigned long begin, end; local 24 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) 26 for (v = begin; v < end; v+=L1_CACHE_BYTES) { 42 unsigned long begin, end; local 45 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) 48 for (v = begin; v < end; v+=L1_CACHE_BYTES) 73 unsigned long begin, end; 76 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) 79 for (v = begin; v < end; v+=L1_CACHE_BYTES)
|
/linux-master/net/ceph/ |
H A D | decode.c | 10 ceph_decode_entity_addr_versioned(void **p, void *end, argument 18 ret = ceph_start_decoding(p, end, 1, "entity_addr_t", &struct_v, 26 ceph_decode_copy_safe(p, end, &addr->type, sizeof(addr->type), bad); 28 ceph_decode_copy_safe(p, end, &addr->nonce, sizeof(addr->nonce), bad); 30 ceph_decode_32_safe(p, end, addr_len, bad); 36 ceph_decode_copy_safe(p, end, &addr->in_addr, addr_len, bad); 50 ceph_decode_entity_addr_legacy(void **p, void *end, argument 56 ceph_decode_skip_n(p, end, 3, bad); 63 ceph_decode_copy_safe(p, end, &addr->nonce, sizeof(addr->nonce), bad); 65 ceph_decode_copy_safe(p, end, 75 ceph_decode_entity_addr(void **p, void *end, struct ceph_entity_addr *addr) argument 95 ceph_decode_entity_addrvec(void **p, void *end, bool msgr2, struct ceph_entity_addr *addr) argument [all...] |
H A D | armor.c | 5 int ceph_armor(char *dst, const char *src, const char *end); 6 int ceph_unarmor(char *dst, const char *src, const char *end); 37 int ceph_armor(char *dst, const char *src, const char *end) argument 42 while (src < end) { 47 if (src < end) { 50 if (src < end) { 75 int ceph_unarmor(char *dst, const char *src, const char *end) argument 79 while (src < end) { 86 if (src + 4 > end)
|
/linux-master/arch/powerpc/mm/book3s32/ |
H A D | tlb.c | 37 * - flush_tlb_range(vma, start, end) flushes a range of pages 38 * - flush_tlb_kernel_range(start, end) flushes kernel pages 50 void hash__flush_range(struct mm_struct *mm, unsigned long start, unsigned long end) argument 58 if (start >= end) 60 end = (end - 1) | ~PAGE_MASK; 64 if (pmd_end > end) 65 pmd_end = end; 70 if (pmd_end == end)
|
/linux-master/tools/testing/selftests/kvm/lib/ |
H A D | guest_sprintf.c | 6 #define APPEND_BUFFER_SAFE(str, end, v) \ 8 GUEST_ASSERT(str < end); \ 43 static char *number(char *str, const char *end, long num, int base, int size, argument 94 APPEND_BUFFER_SAFE(str, end, ' '); 96 APPEND_BUFFER_SAFE(str, end, sign); 99 APPEND_BUFFER_SAFE(str, end, '0'); 101 APPEND_BUFFER_SAFE(str, end, '0'); 102 APPEND_BUFFER_SAFE(str, end, 'x'); 107 APPEND_BUFFER_SAFE(str, end, c); 109 APPEND_BUFFER_SAFE(str, end, ' 120 char *str, *end; local [all...] |
/linux-master/arch/arm64/kernel/ |
H A D | sys_compat.c | 24 __do_compat_cache_op(unsigned long start, unsigned long end) argument 29 unsigned long chunk = min(PAGE_SIZE, end - start); 49 } while (start < end); 55 do_compat_cache_op(unsigned long start, unsigned long end, int flags) argument 57 if (end < start || flags) 60 if (!access_ok((const void __user *)start, end - start)) 63 return __do_compat_cache_op(start, end);
|
/linux-master/arch/csky/abiv2/inc/abi/ |
H A D | cacheflush.h | 16 #define flush_cache_range(vma, start, end) do { } while (0) 37 #define flush_icache_range(start, end) cache_wbinv_range(start, end) 40 unsigned long start, unsigned long end); 43 #define flush_cache_vmap(start, end) do { } while (0) 44 #define flush_cache_vmap_early(start, end) do { } while (0) 45 #define flush_cache_vunmap(start, end) do { } while (0)
|