/linux-master/kernel/bpf/ |
H A D | arena.c | 101 if (attr->key_size || attr->value_size || attr->max_entries == 0 || 112 vm_range = (u64)attr->max_entries * PAGE_SIZE;
|
H A D | arraymap.c | 26 for (i = 0; i < array->map.max_entries; i++) { 37 for (i = 0; i < array->map.max_entries; i++) { 58 if (attr->max_entries == 0 || attr->key_size != 4 || 84 u32 elem_size, index_mask, max_entries; local 91 max_entries = attr->max_entries; 93 /* On 32 bit archs roundup_pow_of_two() with max_entries that has 97 mask64 = fls_long(max_entries - 1); 106 max_entries = index_mask + 1; 108 if (max_entries < att [all...] |
H A D | bloom_filter.c | 102 attr->max_entries == 0 || 127 if (check_mul_overflow(attr->max_entries, nr_hash_funcs, &nr_bits) ||
|
H A D | bpf_local_storage.c | 675 attr->max_entries ||
|
H A D | bpf_struct_ops.c | 887 if (attr->key_size != sizeof(unsigned int) || attr->max_entries != 1 ||
|
H A D | core.c | 2014 if (unlikely(index >= array->map.max_entries))
|
H A D | cpumap.c | 90 if (attr->max_entries == 0 || attr->key_size != 4 || 97 if (attr->max_entries > NR_CPUS) 107 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries * 520 if (key_cpu >= map->max_entries) 541 if (unlikely(key_cpu >= cmap->map.max_entries)) 584 for (i = 0; i < cmap->map.max_entries; i++) { 607 if (key >= map->max_entries) 629 if (index >= cmap->map.max_entries) { 634 if (index == cmap->map.max_entries - 1) 651 usage += (u64)map->max_entries * sizeo [all...] |
H A D | devmap.c | 118 if (attr->max_entries == 0 || attr->key_size != 4 || 136 if (dtab->map.max_entries > 1UL << 31) 139 dtab->n_buckets = roundup_pow_of_two(dtab->map.max_entries); 148 dtab->netdev_map = bpf_map_area_alloc((u64) dtab->map.max_entries * 224 for (i = 0; i < dtab->map.max_entries; i++) { 249 if (index >= dtab->map.max_entries) { 254 if (index == dtab->map.max_entries - 1) 441 if (key >= map->max_entries) 619 for (i = 0; i < map->max_entries; i++) { 737 for (i = 0; i < map->max_entries; [all...] |
H A D | hashtab.c | 226 u32 num_entries = htab->map.max_entries; 245 u32 num_entries = htab->map.max_entries; 279 for (i = 0; i < htab->map.max_entries; i++) { 320 u32 num_entries = htab->map.max_entries; 450 if (attr->max_entries == 0 || attr->key_size == 0 || 495 htab->map.max_entries = roundup(attr->max_entries, 497 if (htab->map.max_entries < attr->max_entries) 498 htab->map.max_entries [all...] |
H A D | local_storage.c | 312 if (attr->max_entries) 313 /* max_entries is not used and enforced to be 0 */
|
H A D | lpm_trie.c | 328 if (trie->n_entries == trie->map.max_entries) { 551 if (attr->max_entries == 0 ||
|
H A D | map_in_map.c | 48 inner_map_meta->max_entries = inner_map->max_entries;
|
H A D | queue_stack_maps.c | 20 u32 size; /* max_entries + 1 */ 49 if (attr->max_entries == 0 || attr->key_size != 0 || 70 size = (u64) attr->max_entries + 1; 264 usage += ((u64)map->max_entries + 1) * map->value_size;
|
H A D | reuseport_array.c | 55 if (unlikely(index >= array->map.max_entries)) 69 if (index >= map->max_entries) 126 for (i = 0; i < map->max_entries; i++) { 155 array = bpf_map_area_alloc(struct_size(array, ptrs, attr->max_entries), numa_node); 246 if (index >= map->max_entries) 323 if (index >= array->map.max_entries) { 328 if (index == array->map.max_entries - 1) 339 return struct_size(array, ptrs, map->max_entries);
|
H A D | ringbuf.c | 194 !is_power_of_2(attr->max_entries) || 195 !PAGE_ALIGNED(attr->max_entries)) 204 rb_map->rb = bpf_ringbuf_alloc(attr->max_entries, rb_map->map.numa_node); 346 nr_data_pages = map->max_entries >> PAGE_SHIFT;
|
H A D | stackmap.c | 51 smap->elems = bpf_map_area_alloc(elem_size * smap->map.max_entries, 61 smap->map.max_entries); 81 if (attr->max_entries == 0 || attr->key_size != 4 || 97 if (attr->max_entries > 1UL << 31) 100 n_buckets = roundup_pow_of_two(attr->max_entries); 671 u64 enties = map->max_entries;
|
H A D | syscall.c | 344 map->max_entries = attr->max_entries; 853 "max_entries:\t%u\n" 862 map->max_entries, 4847 info.max_entries = map->max_entries;
|
H A D | verifier.c | 10048 max = map->max_entries; 11300 * So, in case of global variables, they use array maps with max_entries = 1, 11302 * into the same map value as max_entries is 1, as described above). 15251 WARN_ON_ONCE(map->max_entries != 1); 19846 * if (index >= max_entries) goto out; 19857 map_ptr->max_entries, 2);
|
/linux-master/kernel/bpf/preload/iterators/ |
H A D | iterators.bpf.c | 18 __u32 max_entries; member in struct:bpf_map 89 BPF_SEQ_PRINTF(seq, " id name max_entries cur_entries\n"); 92 map->id, map->name, map->max_entries,
|
/linux-master/kernel/locking/ |
H A D | lockdep.c | 564 int max_entries; local 570 max_entries = MAX_STACK_TRACE_ENTRIES - nr_stack_trace_entries - 573 if (max_entries <= 0) { 582 trace->nr_entries = stack_trace_save(trace->entries, max_entries, 3);
|
/linux-master/kernel/ |
H A D | stacktrace.c | 273 .max_entries = size, 297 .max_entries = size, 321 .max_entries = size, 347 .max_entries = size, 367 .max_entries = size,
|
/linux-master/kernel/trace/ |
H A D | bpf_trace.c | 561 if (unlikely(index >= array->map.max_entries)) 633 if (unlikely(index >= array->map.max_entries)) 806 if (unlikely(idx >= array->map.max_entries))
|
H A D | trace_events_hist.c | 5321 unsigned int max_entries) 5326 for (i = 0; i < max_entries; i++) { 5319 hist_trigger_stacktrace_print(struct seq_file *m, unsigned long *stacktrace_entries, unsigned int max_entries) argument
|
/linux-master/lib/ |
H A D | fw_table.c | 135 * @max_entries: how many entries can we process? 154 int proc_num, unsigned int max_entries) 180 if (!max_entries || count < max_entries) 203 if (max_entries && count > max_entries) { 205 id, proc->id, count - max_entries, count); 150 acpi_parse_entries_array(char *id, unsigned long table_size, union fw_table_header *table_header, unsigned long max_length, struct acpi_subtable_proc *proc, int proc_num, unsigned int max_entries) argument
|
H A D | test_bpf.c | 15501 progs->map.max_entries = ntests + 1;
|