/linux-master/kernel/bpf/ |
H A D | cpumap.c | 93 attr->map_flags & ~BPF_F_NUMA_NODE) 529 u64 map_flags) 539 if (unlikely(map_flags > BPF_EXIST)) 543 if (unlikely(map_flags == BPF_NOEXIST)) 528 cpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument
|
H A D | devmap.c | 121 attr->map_flags & ~DEV_CREATE_FLAG_MASK) 127 attr->map_flags |= BPF_F_RDONLY_PROG; 907 void *key, void *value, u64 map_flags) 914 if (unlikely(map_flags > BPF_EXIST)) 918 if (unlikely(map_flags == BPF_NOEXIST)) 949 u64 map_flags) 952 map, key, value, map_flags); 956 void *key, void *value, u64 map_flags) 968 if (unlikely(map_flags > BPF_EXIST || !val.ifindex)) 974 if (old_dev && (map_flags 906 __dev_map_update_elem(struct net *net, struct bpf_map *map, void *key, void *value, u64 map_flags) argument 948 dev_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 955 __dev_map_hash_update_elem(struct net *net, struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1008 dev_map_hash_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument [all...] |
H A D | hashtab.c | 134 return !(htab->map.map_flags & BPF_F_NO_PREALLOC); 353 htab->map.map_flags & BPF_F_NO_COMMON_LRU, 425 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); 426 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); 427 bool zero_seed = (attr->map_flags & BPF_F_ZERO_SEED); 437 if (attr->map_flags & ~HTAB_CREATE_FLAG_MASK || 438 !bpf_map_flags_access_ok(attr->map_flags)) 480 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); 481 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); 545 if (htab->map.map_flags 1082 check_flags(struct bpf_htab *htab, struct htab_elem *l_old, u64 map_flags) argument 1097 htab_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1200 htab_lru_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1268 __htab_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags, bool onallcpus) argument 1323 __htab_lru_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags, bool onallcpus) argument 1393 htab_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1399 htab_lru_percpu_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 1703 u64 elem_map_flags, map_flags; local 2413 bpf_percpu_hash_update(struct bpf_map *map, void *key, void *value, u64 map_flags) argument 2550 bpf_fd_htab_map_update_elem(struct bpf_map *map, struct file *map_file, void *key, void *value, u64 map_flags) argument [all...] |
H A D | local_storage.c | 212 void *value, u64 map_flags) 219 if (map_flags != BPF_ANY && map_flags != BPF_EXIST) 308 if (attr->map_flags & ~LOCAL_STORAGE_CREATE_FLAG_MASK || 309 !bpf_map_flags_access_ok(attr->map_flags)) 211 bpf_percpu_cgroup_storage_update(struct bpf_map *_map, void *key, void *value, u64 map_flags) argument
|
H A D | lpm_trie.c | 565 !(attr->map_flags & BPF_F_NO_PREALLOC) || 566 attr->map_flags & ~LPM_CREATE_FLAG_MASK || 567 !bpf_map_flags_access_ok(attr->map_flags) ||
|
H A D | map_in_map.c | 47 inner_map_meta->map_flags = inner_map->map_flags; 104 meta0->map_flags == meta1->map_flags &&
|
H A D | queue_stack_maps.c | 51 attr->map_flags & ~QUEUE_STACK_CREATE_FLAG_MASK || 52 !bpf_map_flags_access_ok(attr->map_flags))
|
H A D | reuseport_array.c | 192 u32 map_flags) 194 if (osk && map_flags == BPF_NOEXIST) 197 if (!osk && map_flags == BPF_EXIST) 233 void *value, u64 map_flags) 243 if (map_flags > BPF_EXIST) 273 map_flags); 289 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags); 188 reuseport_array_update_check(const struct reuseport_array *array, const struct sock *nsk, const struct sock *osk, const struct sock_reuseport *nsk_reuse, u32 map_flags) argument 232 bpf_fd_reuseport_array_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument
|
H A D | ringbuf.c | 190 if (attr->map_flags & ~RINGBUF_CREATE_FLAG_MASK)
|
H A D | stackmap.c | 36 return (map->map_flags & BPF_F_STACK_BUILD_ID); 77 if (attr->map_flags & ~STACK_CREATE_FLAG_MASK) 86 if (attr->map_flags & BPF_F_STACK_BUILD_ID) { 631 u64 map_flags) 630 stack_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument
|
H A D | syscall.c | 345 map->map_flags = bpf_map_flags_retain_permanent(attr->map_flags); 866 "map_flags:\t%#x\n" 875 map->map_flags, 951 if (map->map_flags & BPF_F_RDONLY_PROG) { 1109 if (map->map_flags & (BPF_F_RDONLY_PROG | BPF_F_WRONLY_PROG)) { 1214 token_flag = attr->map_flags & BPF_F_TOKEN_FD; 1215 attr->map_flags &= ~BPF_F_TOKEN_FD; 1230 f_flags = bpf_get_file_flag(attr->map_flags); 4870 info.map_flags [all...] |
H A D | verifier.c | 6304 return (map->map_flags & BPF_F_RDONLY_PROG) && 10067 if ((map->map_flags & BPF_F_RDONLY_PROG) && 19829 (((struct bpf_map *)env->prog->aux->arena)->map_flags & BPF_F_NO_USER_CONV)) {
|
/linux-master/mm/ |
H A D | mremap.c | 876 unsigned long map_flags = 0; local 932 map_flags |= MAP_FIXED; 935 map_flags |= MAP_SHARED; 939 map_flags); 1148 unsigned long map_flags = 0; local 1150 map_flags |= MAP_SHARED; 1155 map_flags);
|
/linux-master/net/core/ |
H A D | bpf_sk_storage.c | 98 void *value, u64 map_flags) 109 map_flags, GFP_ATOMIC); 175 if (!(smap->map.map_flags & BPF_F_CLONE)) 97 bpf_fd_sk_storage_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument
|
H A D | sock_map.c | 47 attr->map_flags & ~SOCK_CREATE_FLAG_MASK) 1110 attr->map_flags & ~SOCK_CREATE_FLAG_MASK)
|
/linux-master/net/xdp/ |
H A D | xskmap.c | 72 attr->map_flags & ~(BPF_F_NUMA_NODE | BPF_F_RDONLY | BPF_F_WRONLY)) 162 u64 map_flags) 172 if (unlikely(map_flags > BPF_EXIST)) 200 } else if (old_xs && map_flags == BPF_NOEXIST) { 203 } else if (!old_xs && map_flags == BPF_EXIST) { 161 xsk_map_update_elem(struct bpf_map *map, void *key, void *value, u64 map_flags) argument
|
/linux-master/samples/bpf/ |
H A D | map_perf_test.bpf.c | 36 __uint(map_flags, BPF_F_NO_COMMON_LRU); 44 __uint(map_flags, BPF_F_NUMA_NODE); 70 __uint(map_flags, BPF_F_NO_PREALLOC); 78 __uint(map_flags, BPF_F_NO_PREALLOC); 86 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | map_perf_test_user.c | 137 .map_flags = BPF_F_NUMA_NODE,
|
H A D | tcp_dumpstats_kern.c | 17 __u32 map_flags; member in struct:__anon2008 22 .map_flags = BPF_F_NO_PREALLOC,
|
H A D | test_lru_dist.c | 202 static int create_map(int map_type, int map_flags, unsigned int size) argument 205 .map_flags = map_flags, 307 static void test_parallel_lru_dist(int map_type, int map_flags, argument 313 printf("%s (map_type:%d map_flags:0x%X):\n", __func__, map_type, 314 map_flags); 316 if (map_flags & BPF_F_NO_COMMON_LRU) 317 lru_map_fd = create_map(map_type, map_flags, 320 lru_map_fd = create_map(map_type, map_flags, 332 static void test_lru_loss0(int map_type, int map_flags) argument 386 test_lru_loss1(int map_type, int map_flags) argument 465 test_parallel_lru_loss(int map_type, int map_flags, int nr_tasks) argument 489 int map_flags[] = {0, BPF_F_NO_COMMON_LRU}; local [all...] |
H A D | xdp_router_ipv4.bpf.c | 47 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | xdp_sample.bpf.h | 21 __uint(map_flags, BPF_F_MMAPABLE);
|
/linux-master/tools/bpf/bpftool/ |
H A D | map.c | 465 jsonw_printf(wtr, "%d", info->map_flags); 556 printf("flags 0x%x", info->map_flags); 1286 if (parse_u32_arg(&argc, &argv, &attr.map_flags,
|
/linux-master/tools/bpf/runqslower/ |
H A D | runqslower.bpf.c | 15 __uint(map_flags, BPF_F_NO_PREALLOC);
|
/linux-master/tools/include/uapi/linux/ |
H A D | bpf.h | 1462 __u32 map_flags; /* BPF_MAP_CREATE related member in struct:bpf_attr::__anon3494 1494 * If provided, map_flags should have BPF_F_TOKEN_FD flag set. 6592 __u32 map_flags; member in struct:bpf_map_info
|