/linux-master/samples/bpf/ |
H A D | offwaketime.bpf.c | 31 __uint(max_entries, MAX_ENTRIES); 38 __uint(max_entries, MAX_ENTRIES); 50 __uint(max_entries, MAX_ENTRIES); 57 __uint(max_entries, MAX_ENTRIES);
|
/linux-master/kernel/bpf/preload/iterators/ |
H A D | iterators.bpf.c | 18 __u32 max_entries; member in struct:bpf_map 89 BPF_SEQ_PRINTF(seq, " id name max_entries cur_entries\n"); 92 map->id, map->name, map->max_entries,
|
/linux-master/arch/parisc/net/ |
H A D | bpf_jit_comp64.c | 361 /* max_entries = array->map.max_entries; */ 362 off = offsetof(struct bpf_array, map.max_entries); 363 BUILD_BUG_ON(sizeof(bpfa.map.max_entries) != 4); 367 * if (index >= max_entries)
|
H A D | bpf_jit_comp32.c | 959 /* max_entries = array->map.max_entries; */ 960 off = offsetof(struct bpf_array, map.max_entries); 961 BUILD_BUG_ON(sizeof(bpfa.map.max_entries) != 4); 965 * if (index >= max_entries)
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_sockmap_kern.h | 34 __uint(max_entries, 20); 41 __uint(max_entries, 20); 48 __uint(max_entries, 20); 55 __uint(max_entries, 1); 62 __uint(max_entries, 1); 69 __uint(max_entries, 6); 76 __uint(max_entries, 1); 83 __uint(max_entries, 3); 90 __uint(max_entries, 20);
|
H A D | test_sockmap_pass_prog.c | 7 __uint(max_entries, 20); 14 __uint(max_entries, 20); 21 __uint(max_entries, 20);
|
H A D | test_subprogs_extable.c | 9 __uint(max_entries, 8);
|
H A D | test_sockmap_drop_prog.c | 7 __uint(max_entries, 20); 14 __uint(max_entries, 20); 21 __uint(max_entries, 20);
|
H A D | test_sock_fields.c | 21 __uint(max_entries, __NR_BPF_LINUM_ARRAY_IDX);
|
H A D | sock_destroy_prog.c | 15 __uint(max_entries, 1); 22 __uint(max_entries, 1);
|
H A D | map_kptr.c | 16 __uint(max_entries, 1); 23 __uint(max_entries, 1); 30 __uint(max_entries, 1); 37 __uint(max_entries, 1); 44 __uint(max_entries, 1); 52 __uint(max_entries, 1); 60 __uint(max_entries, 1); 67 __uint(max_entries, 1); 101 __uint(max_entries, 1); \
|
H A D | map_kptr_fail.c | 20 __uint(max_entries, 1);
|
H A D | dynptr_fail.c | 22 __uint(max_entries, 1); 29 __uint(max_entries, 1); 36 __uint(max_entries, 1); 43 __uint(max_entries, 1); 56 __uint(max_entries, 4096);
|
H A D | dynptr_success.c | 25 __uint(max_entries, 4096); 30 __uint(max_entries, 1);
|
H A D | inner_array_lookup.c | 8 __uint(max_entries, 5); 15 __uint(max_entries, 3);
|
H A D | xdpwall.c | 45 __uint(max_entries, 16); 52 __uint(max_entries, 16); 59 __uint(max_entries, 16); 67 __uint(max_entries, 16); 74 __uint(max_entries, 16);
|
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bpf_hashmap_lookup.c | 23 __u32 max_entries; member in struct:__anon5765 29 .max_entries = 1000, 47 { "max_entries", ARG_MAX_ENTRIES, "MAX_ENTRIES", 0, 80 fprintf(stderr, "invalid max_entries"); 83 args.max_entries = ret; 121 if (args.nr_entries > args.max_entries) { 123 args.max_entries, args.nr_entries); 166 bpf_map__set_max_entries(ctx.skel->maps.hash_map_bench, args.max_entries);
|
H A D | bench_bpf_hashmap_full_update.c | 40 int map_fd, i, max_entries; local 60 max_entries = bpf_map__max_entries(ctx.skel->maps.hash_map_bench); 61 for (i = 0; i < max_entries; i++)
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | sample_filter.bpf.c | 15 __uint(max_entries, MAX_FILTERS);
|
/linux-master/tools/lib/bpf/ |
H A D | gen_loader.c | 244 * For example: when ctx->map.max_entries == 0, keep default max_entries from bpf.c 451 __u32 key_size, __u32 value_size, __u32 max_entries, 469 attr.max_entries = max_entries; 491 /* conditionally update max_entries */ 493 move_ctx2blob(gen, attr_field(map_create_attr, max_entries), 4, 496 offsetof(struct bpf_map_desc, max_entries), 497 true /* check that max_entries != 0 */); 448 bpf_gen__map_create(struct bpf_gen *gen, enum bpf_map_type map_type, const char *map_name, __u32 key_size, __u32 value_size, __u32 max_entries, struct bpf_map_create_opts *map_attr, int map_idx) argument
|
/linux-master/net/xdp/ |
H A D | xskmap.c | 70 if (attr->max_entries == 0 || attr->key_size != 4 || 76 size = struct_size(m, xsk_map, attr->max_entries); 92 return struct_size(m, xsk_map, map->max_entries) + 110 if (index >= m->map.max_entries) { 115 if (index == m->map.max_entries - 1) 127 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); 145 if (key >= map->max_entries) 174 if (unlikely(i >= m->map.max_entries)) 229 if (k >= map->max_entries) 262 return meta0->max_entries [all...] |
/linux-master/kernel/bpf/ |
H A D | reuseport_array.c | 55 if (unlikely(index >= array->map.max_entries)) 69 if (index >= map->max_entries) 126 for (i = 0; i < map->max_entries; i++) { 155 array = bpf_map_area_alloc(struct_size(array, ptrs, attr->max_entries), numa_node); 246 if (index >= map->max_entries) 323 if (index >= array->map.max_entries) { 328 if (index == array->map.max_entries - 1) 339 return struct_size(array, ptrs, map->max_entries);
|
H A D | local_storage.c | 312 if (attr->max_entries) 313 /* max_entries is not used and enforced to be 0 */
|
/linux-master/fs/ext4/ |
H A D | migrate.c | 96 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; local 103 for (i = 0; i < max_entries; i++) { 125 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; local 132 for (i = 0; i < max_entries; i++) { 140 lb->curr_block += max_entries; 155 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; local 162 for (i = 0; i < max_entries; i++) { 170 lb->curr_block += max_entries * max_entries; 185 unsigned long max_entries local 225 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; local 417 unsigned long max_entries; local [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | test_tcp_hdr_options.h | 91 __uint(max_entries, 2);
|