/linux-master/tools/perf/util/bpf_skel/ |
H A D | augmented_raw_syscalls.bpf.c | 30 __uint(max_entries, MAX_CPUS); 42 __uint(max_entries, 512); 54 __uint(max_entries, 512); 79 __uint(max_entries, 64); 124 __uint(max_entries, 1);
|
H A D | bperf_cgroup.bpf.c | 20 __uint(max_entries, 1); 28 __uint(max_entries, 1);
|
H A D | bperf_follower.bpf.c | 12 __uint(max_entries, 1); 19 __uint(max_entries, 1);
|
H A D | bperf_leader.bpf.c | 18 __uint(max_entries, 1); 25 __uint(max_entries, 1);
|
H A D | bpf_prog_profiler.bpf.c | 19 __uint(max_entries, 1); 27 __uint(max_entries, 1);
|
H A D | func_latency.bpf.c | 14 __uint(max_entries, 10000); 21 __uint(max_entries, 1); 28 __uint(max_entries, 1); 35 __uint(max_entries, NUM_BUCKET);
|
H A D | kwork_top.bpf.c | 61 __uint(max_entries, MAX_ENTRIES); 68 __uint(max_entries, MAX_ENTRIES); 75 __uint(max_entries, MAX_ENTRIES); 82 __uint(max_entries, MAX_NR_CPUS);
|
H A D | kwork_trace.bpf.c | 39 __uint(max_entries, KWORK_COUNT); 46 __uint(max_entries, KWORK_COUNT); 53 __uint(max_entries, KWORK_COUNT); 60 __uint(max_entries, 1); 67 __uint(max_entries, 1);
|
H A D | lock_contention.bpf.c | 27 __uint(max_entries, MAX_ENTRIES); 35 __uint(max_entries, MAX_ENTRIES); 43 __uint(max_entries, 1); 51 __uint(max_entries, MAX_ENTRIES); 58 __uint(max_entries, MAX_ENTRIES); 65 __uint(max_entries, MAX_ENTRIES); 72 __uint(max_entries, 1); 79 __uint(max_entries, 1); 86 __uint(max_entries, 1); 93 __uint(max_entries, [all...] |
H A D | off_cpu.bpf.c | 39 __uint(max_entries, MAX_ENTRIES); 53 __uint(max_entries, MAX_ENTRIES); 60 __uint(max_entries, 1); 67 __uint(max_entries, 1); 74 __uint(max_entries, 1);
|
H A D | sample_filter.bpf.c | 15 __uint(max_entries, MAX_FILTERS);
|
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bpf_hashmap_full_update.c | 40 int map_fd, i, max_entries; local 60 max_entries = bpf_map__max_entries(ctx.skel->maps.hash_map_bench); 61 for (i = 0; i < max_entries; i++)
|
H A D | bench_bpf_hashmap_lookup.c | 23 __u32 max_entries; member in struct:__anon5765 29 .max_entries = 1000, 47 { "max_entries", ARG_MAX_ENTRIES, "MAX_ENTRIES", 0, 80 fprintf(stderr, "invalid max_entries"); 83 args.max_entries = ret; 121 if (args.nr_entries > args.max_entries) { 123 args.max_entries, args.nr_entries); 166 bpf_map__set_max_entries(ctx.skel->maps.hash_map_bench, args.max_entries);
|
/linux-master/tools/testing/selftests/bpf/map_tests/ |
H A D | array_map_batch_ops.c | 15 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, argument 25 for (i = 0; i < max_entries; i++) { 36 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); 40 static void map_batch_verify(int *visited, __u32 max_entries, int *keys, argument 46 memset(visited, 0, max_entries * sizeof(*visited)); 47 for (i = 0; i < max_entries; i++) { 64 for (i = 0; i < max_entries; i++) { 74 const __u32 max_entries = 10; local 84 "array_map", sizeof(int), sizeof(__s64), max_entries, NULL); 92 keys = calloc(max_entries, sizeo [all...] |
H A D | htab_map_batch_ops.c | 14 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, argument 28 for (i = 0; i < max_entries; i++) { 37 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); 41 static void map_batch_verify(int *visited, __u32 max_entries, argument 51 memset(visited, 0, max_entries * sizeof(*visited)); 52 for (i = 0; i < max_entries; i++) { 71 for (i = 0; i < max_entries; i++) { 82 const __u32 max_entries = 10; local 83 value pcpu_values[max_entries]; 93 "hash_map", sizeof(int), sizeof(int), max_entries, NUL [all...] |
H A D | lpm_trie_map_batch_ops.c | 22 static void map_batch_update(int map_fd, __u32 max_entries, argument 33 for (i = 0; i < max_entries; i++) { 40 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); 44 static void map_batch_verify(int *visited, __u32 max_entries, argument 51 memset(visited, 0, max_entries * sizeof(*visited)); 52 for (i = 0; i < max_entries; i++) { 60 for (i = 0; i < max_entries; i++) { 72 const __u32 max_entries = 10; local 82 max_entries, &create_opts); 86 keys = malloc(max_entries * sizeo [all...] |
H A D | map_in_map_batch_ops.c | 127 int err, max_entries = OUTER_MAP_ENTRIES; local 131 fetched_keys = calloc(max_entries, value_size); 132 fetched_values = calloc(max_entries, value_size); 138 step_size <= max_entries; 171 CHECK((total_fetched != max_entries), 173 "total_fetched(%d) and max_entries(%d) error: (%d):%s\n", 174 total_fetched, max_entries, errno, strerror(errno)); 190 __u32 max_entries = OUTER_MAP_ENTRIES; local 197 outer_map_keys = calloc(max_entries, value_size); 198 inner_map_fds = calloc(max_entries, value_siz [all...] |
H A D | map_percpu_stats.c | 300 opts.n = info.max_entries; 340 int max_entries; local 344 max_entries = MAX_ENTRIES_HASH_OF_MAPS; 346 max_entries = MAX_ENTRIES; 348 map_fd = bpf_map_create(type, name, key_size, val_size, max_entries, map_opts);
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | bpf_obj_id.c | 89 !ASSERT_EQ(map_infos[i].max_entries, 1, "max_entries") ||
|
H A D | btf.c | 72 __u32 max_entries; member in struct:btf_raw_test 140 .max_entries = 4, 195 .max_entries = 4, 220 .max_entries = 4, 261 .max_entries = 4, 306 .max_entries = 1, 328 .max_entries = 1, 350 .max_entries = 1, 372 .max_entries = 1, 397 .max_entries [all...] |
H A D | for_each.c | 13 int i, err, max_entries; local 29 max_entries = bpf_map__max_entries(skel->maps.hashmap); 30 for (i = 0; i < max_entries; i++) { 60 ASSERT_EQ(skel->bss->hashmap_elems, max_entries, "hashmap_elems"); 79 __u32 key, num_cpus, max_entries; local 96 max_entries = bpf_map__max_entries(skel->maps.arraymap); 97 for (i = 0; i < max_entries; i++) { 101 if (i != max_entries - 1)
|
H A D | map_ptr.c | 27 skel->maps.m_ringbuf.max_entries = page_size;
|
H A D | obj_name.c | 57 attr.max_entries = 1;
|
H A D | ringbuf.c | 102 skel->maps.ringbuf.max_entries = page_size; 356 skel_map_key->maps.ringbuf.max_entries = getpagesize();
|
H A D | sockmap_basic.c | 61 __u32 i, max_entries = bpf_map__max_entries(src); local 67 for (i = 0; i < max_entries; i++) {
|