/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_bpf_sk_storage_map.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | task_storage_nodeadlock.c | 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | udp_limit.c | 11 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | nested_trust_failure.c | 15 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | sk_storage_omem_uncharge.c | 17 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | arena_htab.c | 11 __uint(map_flags, BPF_F_MMAPABLE);
|
H A D | test_sk_storage_tracing.c | 17 __uint(map_flags, BPF_F_NO_PREALLOC); 25 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | cgrp_ls_recursion.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC); 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | cgrp_ls_tp_btf.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC); 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | test_skmsg_load_helpers.c | 22 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | socket_cookie_prog.c | 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | tcp_rtt.c | 17 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | test_stacktrace_build_id.c | 30 __uint(map_flags, BPF_F_STACK_BUILD_ID);
|
H A D | arena_list.c | 11 __uint(map_flags, BPF_F_MMAPABLE);
|
H A D | verifier_arena_large.c | 15 __uint(map_flags, BPF_F_MMAPABLE);
|
/linux-master/kernel/bpf/ |
H A D | map_in_map.c | 47 inner_map_meta->map_flags = inner_map->map_flags; 104 meta0->map_flags == meta1->map_flags &&
|
H A D | bpf_local_storage.c | 435 u64 map_flags) 437 if (old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_NOEXIST) 441 if (!old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) 527 void *value, u64 map_flags, gfp_t gfp_flags) 536 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST) || 538 unlikely((map_flags & BPF_F_LOCK) && 542 if (gfp_flags == GFP_KERNEL && (map_flags & ~BPF_F_LOCK) != BPF_NOEXIST) 549 err = check_flags(NULL, map_flags); 567 if ((map_flags & BPF_F_LOCK) && !(map_flags 434 check_flags(const struct bpf_local_storage_data *old_sdata, u64 map_flags) argument 526 bpf_local_storage_update(void *owner, struct bpf_local_storage_map *smap, void *value, u64 map_flags, gfp_t gfp_flags) argument [all...] |
H A D | arraymap.c | 60 attr->map_flags & ~ARRAY_CREATE_FLAG_MASK || 61 !bpf_map_flags_access_ok(attr->map_flags) || 66 attr->map_flags & (BPF_F_MMAPABLE | BPF_F_INNER_MAP)) 70 attr->map_flags & BPF_F_PRESERVE_ELEMS) 119 if (attr->map_flags & BPF_F_MMAPABLE) { 128 if (attr->map_flags & BPF_F_MMAPABLE) { 214 if (map->map_flags & BPF_F_INNER_MAP) 311 u64 map_flags) 317 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST)) 325 if (unlikely(map_flags 310 array_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 349 bpf_percpu_array_update(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 842 bpf_fd_array_map_update_elem(struct bpf_map *map, struct file *map_file, void *key, void *value, u64 map_flags) argument [all...] |
H A D | hashtab.c | 134 return !(htab->map.map_flags & BPF_F_NO_PREALLOC); 350 htab->map.map_flags & BPF_F_NO_COMMON_LRU, 422 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); 423 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); 424 bool zero_seed = (attr->map_flags & BPF_F_ZERO_SEED); 434 if (attr->map_flags & ~HTAB_CREATE_FLAG_MASK || 435 !bpf_map_flags_access_ok(attr->map_flags)) 477 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); 478 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); 542 if (htab->map.map_flags 1079 check_flags(struct bpf_htab *htab, struct htab_elem *l_old, u64 map_flags) argument 1094 htab_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1197 htab_lru_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1265 __htab_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags, bool onallcpus) argument 1320 __htab_lru_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags, bool onallcpus) argument 1390 htab_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1396 htab_lru_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1694 u64 elem_map_flags, map_flags; local 2384 bpf_percpu_hash_update(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 2520 bpf_fd_htab_map_update_elem(struct bpf_map *map, struct file *map_file, void *key, void *value, u64 map_flags) argument [all...] |
/linux-master/tools/testing/selftests/bpf/map_tests/ |
H A D | map_percpu_stats.c | 287 "n_iter=%u, expected %u (map_type=%s,map_flags=%08x)\n", 288 n_iter, n_real, map_type_to_s(info->type), info->map_flags); 315 opts.retry_for_nomem = is_percpu(opts.map_type) && (info.map_flags & BPF_F_NO_PREALLOC); 362 LIBBPF_OPTS(bpf_map_create_opts, map_opts, .map_flags = BPF_F_NO_PREALLOC); 369 LIBBPF_OPTS(bpf_map_create_opts, map_opts, .map_flags = BPF_F_NO_PREALLOC); 384 static int create_lru_hash(__u32 type, __u32 map_flags) argument 386 LIBBPF_OPTS(bpf_map_create_opts, map_opts, .map_flags = map_flags); 394 .map_flags = BPF_F_NO_PREALLOC,
|
/linux-master/samples/bpf/ |
H A D | map_perf_test.bpf.c | 36 __uint(map_flags, BPF_F_NO_COMMON_LRU); 44 __uint(map_flags, BPF_F_NUMA_NODE); 70 __uint(map_flags, BPF_F_NO_PREALLOC); 78 __uint(map_flags, BPF_F_NO_PREALLOC); 86 __uint(map_flags, BPF_F_NO_PREALLOC);
|
/linux-master/include/rdma/ |
H A D | iw_portmap.h | 60 u32 map_flags);
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | bperf_leader.bpf.c | 11 __uint(map_flags, BPF_F_PRESERVE_ELEMS);
|
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bpf_hashmap_lookup.c | 22 __u32 map_flags; member in struct:__anon5765 28 .map_flags = 0, 45 { "map_flags", ARG_MAP_FLAGS, "MAP_FLAGS", 0, 72 fprintf(stderr, "invalid map_flags"); 75 args.map_flags = ret; 169 bpf_map__set_map_flags(ctx.skel->maps.hash_map_bench, args.map_flags);
|
/linux-master/fs/bcachefs/ |
H A D | fs-ioctl.h | 45 #define map_flags(_map, _in) \ macro
|