Searched refs:max_entries (Results 101 - 125 of 417) sorted by path

1234567891011>>

/linux-master/kernel/bpf/
H A Darena.c101 if (attr->key_size || attr->value_size || attr->max_entries == 0 ||
112 vm_range = (u64)attr->max_entries * PAGE_SIZE;
H A Darraymap.c26 for (i = 0; i < array->map.max_entries; i++) {
37 for (i = 0; i < array->map.max_entries; i++) {
58 if (attr->max_entries == 0 || attr->key_size != 4 ||
84 u32 elem_size, index_mask, max_entries; local
91 max_entries = attr->max_entries;
93 /* On 32 bit archs roundup_pow_of_two() with max_entries that has
97 mask64 = fls_long(max_entries - 1);
106 max_entries = index_mask + 1;
108 if (max_entries < att
[all...]
H A Dbloom_filter.c102 attr->max_entries == 0 ||
127 if (check_mul_overflow(attr->max_entries, nr_hash_funcs, &nr_bits) ||
H A Dbpf_local_storage.c675 attr->max_entries ||
H A Dbpf_struct_ops.c887 if (attr->key_size != sizeof(unsigned int) || attr->max_entries != 1 ||
H A Dcore.c2014 if (unlikely(index >= array->map.max_entries))
H A Dcpumap.c90 if (attr->max_entries == 0 || attr->key_size != 4 ||
97 if (attr->max_entries > NR_CPUS)
107 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries *
520 if (key_cpu >= map->max_entries)
541 if (unlikely(key_cpu >= cmap->map.max_entries))
584 for (i = 0; i < cmap->map.max_entries; i++) {
607 if (key >= map->max_entries)
629 if (index >= cmap->map.max_entries) {
634 if (index == cmap->map.max_entries - 1)
651 usage += (u64)map->max_entries * sizeo
[all...]
H A Ddevmap.c118 if (attr->max_entries == 0 || attr->key_size != 4 ||
136 if (dtab->map.max_entries > 1UL << 31)
139 dtab->n_buckets = roundup_pow_of_two(dtab->map.max_entries);
148 dtab->netdev_map = bpf_map_area_alloc((u64) dtab->map.max_entries *
224 for (i = 0; i < dtab->map.max_entries; i++) {
249 if (index >= dtab->map.max_entries) {
254 if (index == dtab->map.max_entries - 1)
441 if (key >= map->max_entries)
619 for (i = 0; i < map->max_entries; i++) {
737 for (i = 0; i < map->max_entries;
[all...]
H A Dhashtab.c226 u32 num_entries = htab->map.max_entries;
245 u32 num_entries = htab->map.max_entries;
279 for (i = 0; i < htab->map.max_entries; i++) {
320 u32 num_entries = htab->map.max_entries;
450 if (attr->max_entries == 0 || attr->key_size == 0 ||
495 htab->map.max_entries = roundup(attr->max_entries,
497 if (htab->map.max_entries < attr->max_entries)
498 htab->map.max_entries
[all...]
H A Dlocal_storage.c312 if (attr->max_entries)
313 /* max_entries is not used and enforced to be 0 */
H A Dlpm_trie.c328 if (trie->n_entries == trie->map.max_entries) {
551 if (attr->max_entries == 0 ||
H A Dmap_in_map.c48 inner_map_meta->max_entries = inner_map->max_entries;
H A Dqueue_stack_maps.c20 u32 size; /* max_entries + 1 */
49 if (attr->max_entries == 0 || attr->key_size != 0 ||
70 size = (u64) attr->max_entries + 1;
264 usage += ((u64)map->max_entries + 1) * map->value_size;
H A Dreuseport_array.c55 if (unlikely(index >= array->map.max_entries))
69 if (index >= map->max_entries)
126 for (i = 0; i < map->max_entries; i++) {
155 array = bpf_map_area_alloc(struct_size(array, ptrs, attr->max_entries), numa_node);
246 if (index >= map->max_entries)
323 if (index >= array->map.max_entries) {
328 if (index == array->map.max_entries - 1)
339 return struct_size(array, ptrs, map->max_entries);
H A Dringbuf.c194 !is_power_of_2(attr->max_entries) ||
195 !PAGE_ALIGNED(attr->max_entries))
204 rb_map->rb = bpf_ringbuf_alloc(attr->max_entries, rb_map->map.numa_node);
346 nr_data_pages = map->max_entries >> PAGE_SHIFT;
H A Dstackmap.c51 smap->elems = bpf_map_area_alloc(elem_size * smap->map.max_entries,
61 smap->map.max_entries);
81 if (attr->max_entries == 0 || attr->key_size != 4 ||
97 if (attr->max_entries > 1UL << 31)
100 n_buckets = roundup_pow_of_two(attr->max_entries);
671 u64 enties = map->max_entries;
H A Dsyscall.c344 map->max_entries = attr->max_entries;
853 "max_entries:\t%u\n"
862 map->max_entries,
4847 info.max_entries = map->max_entries;
H A Dverifier.c10048 max = map->max_entries;
11300 * So, in case of global variables, they use array maps with max_entries = 1,
11302 * into the same map value as max_entries is 1, as described above).
15251 WARN_ON_ONCE(map->max_entries != 1);
19846 * if (index >= max_entries) goto out;
19857 map_ptr->max_entries, 2);
/linux-master/kernel/bpf/preload/iterators/
H A Diterators.bpf.c18 __u32 max_entries; member in struct:bpf_map
89 BPF_SEQ_PRINTF(seq, " id name max_entries cur_entries\n");
92 map->id, map->name, map->max_entries,
/linux-master/kernel/locking/
H A Dlockdep.c564 int max_entries; local
570 max_entries = MAX_STACK_TRACE_ENTRIES - nr_stack_trace_entries -
573 if (max_entries <= 0) {
582 trace->nr_entries = stack_trace_save(trace->entries, max_entries, 3);
/linux-master/kernel/
H A Dstacktrace.c273 .max_entries = size,
297 .max_entries = size,
321 .max_entries = size,
347 .max_entries = size,
367 .max_entries = size,
/linux-master/kernel/trace/
H A Dbpf_trace.c561 if (unlikely(index >= array->map.max_entries))
633 if (unlikely(index >= array->map.max_entries))
806 if (unlikely(idx >= array->map.max_entries))
H A Dtrace_events_hist.c5321 unsigned int max_entries)
5326 for (i = 0; i < max_entries; i++) {
5319 hist_trigger_stacktrace_print(struct seq_file *m, unsigned long *stacktrace_entries, unsigned int max_entries) argument
/linux-master/lib/
H A Dfw_table.c135 * @max_entries: how many entries can we process?
154 int proc_num, unsigned int max_entries)
180 if (!max_entries || count < max_entries)
203 if (max_entries && count > max_entries) {
205 id, proc->id, count - max_entries, count);
150 acpi_parse_entries_array(char *id, unsigned long table_size, union fw_table_header *table_header, unsigned long max_length, struct acpi_subtable_proc *proc, int proc_num, unsigned int max_entries) argument
H A Dtest_bpf.c15501 progs->map.max_entries = ntests + 1;

Completed in 766 milliseconds

1234567891011>>