/linux-master/net/core/ |
H A D | filter.c | 4879 if (unlikely(idx >= array->map.max_entries))
|
H A D | sock_map.c | 35 if (attr->max_entries == 0 || 49 stab->sks = bpf_map_area_alloc((u64) stab->map.max_entries * 342 for (i = 0; i < stab->map.max_entries; i++) { 376 if (unlikely(key >= map->max_entries)) 445 if (unlikely(i >= map->max_entries)) 458 if (i == stab->map.max_entries - 1) 460 if (i >= stab->map.max_entries) 479 if (unlikely(idx >= map->max_entries)) 713 if (unlikely(info->index >= info->map->max_entries)) 807 usage += (u64)map->max_entries * sizeo [all...] |
/linux-master/net/xdp/ |
H A D | xsk.c | 480 static u32 xsk_tx_peek_release_fallback(struct xsk_buff_pool *pool, u32 max_entries) argument 485 while (nb_pkts < max_entries && xsk_tx_peek_desc(pool, &descs[nb_pkts]))
|
H A D | xskmap.c | 70 if (attr->max_entries == 0 || attr->key_size != 4 || 76 size = struct_size(m, xsk_map, attr->max_entries); 92 return struct_size(m, xsk_map, map->max_entries) + 110 if (index >= m->map.max_entries) { 115 if (index == m->map.max_entries - 1) 127 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); 145 if (key >= map->max_entries) 174 if (unlikely(i >= m->map.max_entries)) 229 if (k >= map->max_entries) 262 return meta0->max_entries [all...] |
/linux-master/samples/bpf/ |
H A D | cpustat_kern.c | 58 __uint(max_entries, MAX_CPU * MAP_OFF_NUM); 66 __uint(max_entries, MAX_CPU * MAX_CSTATE_ENTRIES); 74 __uint(max_entries, MAX_CPU * MAX_PSTATE_ENTRIES);
|
H A D | hbm_kern.h | 68 __uint(max_entries, 1);
|
H A D | ibumad_kern.c | 23 __uint(max_entries, 256); /* Room for all Classes */ 30 __uint(max_entries, 256); /* Room for all Classes */
|
H A D | lathist_kern.c | 25 __uint(max_entries, MAX_CPU); 68 __uint(max_entries, MAX_CPU * MAX_ENTRIES);
|
H A D | lwt_len_hist.bpf.c | 21 __uint(max_entries, 1024);
|
H A D | map_perf_test.bpf.c | 21 __uint(max_entries, MAX_ENTRIES); 28 __uint(max_entries, 10000); 35 __uint(max_entries, 10000); 43 __uint(max_entries, MAX_ENTRIES); 50 __uint(max_entries, MAX_NR_CPUS); 62 __uint(max_entries, MAX_ENTRIES); 69 __uint(max_entries, MAX_ENTRIES); 77 __uint(max_entries, MAX_ENTRIES); 85 __uint(max_entries, 10000); 93 __uint(max_entries, MAX_ENTRIE [all...] |
H A D | offwaketime.bpf.c | 31 __uint(max_entries, MAX_ENTRIES); 38 __uint(max_entries, MAX_ENTRIES); 50 __uint(max_entries, MAX_ENTRIES); 57 __uint(max_entries, MAX_ENTRIES);
|
H A D | sampleip_kern.c | 19 __uint(max_entries, MAX_IPS);
|
H A D | sockex1_kern.c | 12 __uint(max_entries, 256);
|
H A D | sockex2_kern.c | 197 __uint(max_entries, 1024);
|
H A D | sockex3_kern.c | 66 __uint(max_entries, 32); 87 __uint(max_entries, 1024); 254 __uint(max_entries, 8);
|
H A D | spintest.bpf.c | 20 __uint(max_entries, 1024); 26 __uint(max_entries, 1024); 33 __uint(max_entries, 10000);
|
H A D | syscall_tp_kern.c | 36 __uint(max_entries, 1); 43 __uint(max_entries, 1);
|
H A D | test_cgrp2_tc.bpf.c | 24 __uint(max_entries, 1);
|
H A D | test_current_task_under_cgroup.bpf.c | 18 __uint(max_entries, 1); 25 __uint(max_entries, 1);
|
H A D | test_map_in_map.bpf.c | 25 __uint(max_entries, MAX_NR_PORTS); 33 __uint(max_entries, 1); 41 __uint(max_entries, 1); 49 __uint(max_entries, 1); 55 __uint(max_entries, MAX_NR_PORTS); 63 __uint(max_entries, 1); 71 __uint(max_entries, 1);
|
H A D | test_probe_write_user.bpf.c | 18 __uint(max_entries, 256);
|
H A D | trace_event_kern.c | 24 __uint(max_entries, 10000); 31 __uint(max_entries, 10000);
|
H A D | trace_output.bpf.c | 9 __uint(max_entries, 2);
|
H A D | tracex2.bpf.c | 17 __uint(max_entries, 1024); 76 __uint(max_entries, 1024);
|
H A D | tracex3.bpf.c | 22 __uint(max_entries, 4096); 54 __uint(max_entries, SLOTS);
|