/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | ima.c | 16 __uint(max_entries, 1 << 12);
|
H A D | test_skeleton.c | 67 __uint(max_entries, 1);
|
H A D | test_global_func9.c | 18 __uint(max_entries, 1);
|
H A D | xdp_features.c | 37 __uint(max_entries, 1); 44 __uint(max_entries, 1); 51 __uint(max_entries, 1); 58 __uint(max_entries, 1);
|
H A D | verifier_jeq_infer_not_null.c | 10 __uint(max_entries, 1);
|
H A D | task_kfunc_common.h | 20 __uint(max_entries, 1);
|
H A D | tailcall_bpf2bpf3.c | 9 __uint(max_entries, 2);
|
H A D | local_kptr_stash_fail.c | 39 __uint(max_entries, 2);
|
H A D | preempted_bpf_ma_op.c | 22 __uint(max_entries, 2048);
|
H A D | xsk_xdp_progs.c | 11 __uint(max_entries, 2);
|
H A D | test_verify_pkcs7_sig.c | 31 __uint(max_entries, 1);
|
H A D | verifier_value_illegal_alu.c | 17 __uint(max_entries, 1);
|
H A D | async_stack_depth.c | 13 __uint(max_entries, 64);
|
H A D | arena_list.c | 12 __uint(max_entries, 100); /* number of pages */
|
H A D | xdpwall.c | 45 __uint(max_entries, 16); 52 __uint(max_entries, 16); 59 __uint(max_entries, 16); 67 __uint(max_entries, 16); 74 __uint(max_entries, 16);
|
H A D | timer.c | 18 __uint(max_entries, 1000); 26 __uint(max_entries, 1000); 37 __uint(max_entries, 2); 44 __uint(max_entries, 4); 51 __uint(max_entries, 1); 101 * should be larger than ~ lru->max_entries * 2
|
/linux-master/arch/um/kernel/ |
H A D | stacktrace.c | 50 if (trace->nr_entries >= trace->max_entries)
|
/linux-master/samples/bpf/ |
H A D | lwt_len_hist.bpf.c | 21 __uint(max_entries, 1024);
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | kwork_top.bpf.c | 61 __uint(max_entries, MAX_ENTRIES); 68 __uint(max_entries, MAX_ENTRIES); 75 __uint(max_entries, MAX_ENTRIES); 82 __uint(max_entries, MAX_NR_CPUS);
|
H A D | kwork_trace.bpf.c | 39 __uint(max_entries, KWORK_COUNT); 46 __uint(max_entries, KWORK_COUNT); 53 __uint(max_entries, KWORK_COUNT); 60 __uint(max_entries, 1); 67 __uint(max_entries, 1);
|
H A D | augmented_raw_syscalls.bpf.c | 30 __uint(max_entries, MAX_CPUS); 42 __uint(max_entries, 512); 54 __uint(max_entries, 512); 79 __uint(max_entries, 64); 124 __uint(max_entries, 1);
|
/linux-master/kernel/bpf/ |
H A D | cpumap.c | 90 if (attr->max_entries == 0 || attr->key_size != 4 || 97 if (attr->max_entries > NR_CPUS) 107 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries * 520 if (key_cpu >= map->max_entries) 541 if (unlikely(key_cpu >= cmap->map.max_entries)) 584 for (i = 0; i < cmap->map.max_entries; i++) { 607 if (key >= map->max_entries) 629 if (index >= cmap->map.max_entries) { 634 if (index == cmap->map.max_entries - 1) 651 usage += (u64)map->max_entries * sizeo [all...] |
H A D | devmap.c | 118 if (attr->max_entries == 0 || attr->key_size != 4 || 136 if (dtab->map.max_entries > 1UL << 31) 139 dtab->n_buckets = roundup_pow_of_two(dtab->map.max_entries); 148 dtab->netdev_map = bpf_map_area_alloc((u64) dtab->map.max_entries * 224 for (i = 0; i < dtab->map.max_entries; i++) { 249 if (index >= dtab->map.max_entries) { 254 if (index == dtab->map.max_entries - 1) 441 if (key >= map->max_entries) 619 for (i = 0; i < map->max_entries; i++) { 737 for (i = 0; i < map->max_entries; [all...] |
/linux-master/tools/testing/selftests/bpf/map_tests/ |
H A D | map_percpu_stats.c | 300 opts.n = info.max_entries; 340 int max_entries; local 344 max_entries = MAX_ENTRIES_HASH_OF_MAPS; 346 max_entries = MAX_ENTRIES; 348 map_fd = bpf_map_create(type, name, key_size, val_size, max_entries, map_opts);
|
/linux-master/drivers/iommu/ |
H A D | io-pgtable-dart.c | 243 int ret = 0, tbl, num_entries, max_entries, map_idx_start; local 282 max_entries = DART_PTES_PER_TABLE(data) - map_idx_start; 283 num_entries = min_t(int, pgcount, max_entries); 304 int i = 0, num_entries, max_entries, unmap_idx_start; local 319 max_entries = DART_PTES_PER_TABLE(data) - unmap_idx_start; 320 num_entries = min_t(int, pgcount, max_entries);
|