/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | tailcall_bpf2bpf6.c | 12 __uint(value_size, sizeof(__u32));
|
H A D | test_queue_stack_map.h | 16 __uint(value_size, sizeof(__u32)); 24 __uint(value_size, sizeof(__u32));
|
H A D | tailcall_bpf2bpf4.c | 9 __uint(value_size, sizeof(__u32)); 16 __uint(value_size, sizeof(__u32));
|
H A D | tailcall2.c | 10 __uint(value_size, sizeof(__u32));
|
H A D | tailcall1.c | 10 __uint(value_size, sizeof(__u32));
|
H A D | test_xdp_with_devmap_helpers.c | 8 __uint(value_size, sizeof(struct bpf_devmap_val));
|
H A D | xdp_redirect_multi_kern.c | 18 __uint(value_size, sizeof(int)); 25 __uint(value_size, sizeof(struct bpf_devmap_val));
|
/linux-master/tools/bpf/bpftool/skeleton/ |
H A D | profiler.bpf.c | 17 __uint(value_size, sizeof(int)); 24 __uint(value_size, sizeof(struct bpf_perf_event_value___local)); 31 __uint(value_size, sizeof(struct bpf_perf_event_value___local)); 38 __uint(value_size, sizeof(u64));
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | btf.c | 69 __u32 value_size; member in struct:btf_raw_test 137 .value_size = 180, 192 .value_size = 68, 217 .value_size = 16, 258 .value_size = 48, 303 .value_size = 48, 325 .value_size = 4, 347 .value_size = 4, 369 .value_size = 4, 394 .value_size [all...] |
/linux-master/samples/bpf/ |
H A D | spintest.bpf.c | 25 __uint(value_size, sizeof(long)); 32 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
H A D | test_current_task_under_cgroup.bpf.c | 17 __uint(value_size, sizeof(u32));
|
/linux-master/kernel/bpf/ |
H A D | bloom_filter.c | 24 u32 value_size, u32 index) 28 if (likely(value_size % 4 == 0)) 29 h = jhash2(value, value_size / 4, bloom->hash_seed + index); 31 h = jhash(value, value_size, bloom->hash_seed + index); 43 h = hash(bloom, value, map->value_size, i); 61 h = hash(bloom, value, map->value_size, i); 86 if (attr->value_size > KMALLOC_MAX_SIZE) 87 /* if value_size is bigger, the user space won't be able to 101 if (attr->key_size != 0 || attr->value_size == 0 || 23 hash(struct bpf_bloom_filter *bloom, void *value, u32 value_size, u32 index) argument
|
H A D | map_in_map.c | 46 inner_map_meta->value_size = inner_map->value_size; 103 meta0->value_size == meta1->value_size &&
|
H A D | map_iter.c | 104 u32 key_acc_size, value_acc_size, key_size, value_size; local 129 value_size = map->value_size; 131 value_size = round_up(map->value_size, 8) * num_possible_cpus(); 133 if (key_acc_size > key_size || value_acc_size > value_size) {
|
H A D | stackmap.c | 48 (u64)smap->map.value_size; 72 u32 value_size = attr->value_size; local 82 value_size < 8 || value_size % 8) 87 if (value_size % sizeof(struct bpf_stack_build_id) || 88 value_size / sizeof(struct bpf_stack_build_id) 91 } else if (value_size / 8 > sysctl_perf_event_max_stack) 286 u32 max_depth = map->value_size / stack_map_data_size(map); 593 memset(value + trace_len, 0, map->value_size 669 u64 value_size = map->value_size; local [all...] |
/linux-master/tools/testing/selftests/bpf/map_tests/ |
H A D | htab_map_batch_ops.c | 84 int err, step, value_size; local 97 value_size = is_pcpu ? sizeof(value) : sizeof(int); 125 memset(values, 0, max_entries * value_size); 144 memset(values, 0, max_entries * value_size); 156 total * value_size, 210 memset(values, 0, max_entries * value_size); 219 total * value_size,
|
H A D | map_in_map_batch_ops.c | 128 __u32 value_size = sizeof(__u32); local 131 fetched_keys = calloc(max_entries, value_size); 132 fetched_values = calloc(max_entries, value_size); 192 __u32 value_size = sizeof(__u32); local 197 outer_map_keys = calloc(max_entries, value_size); 198 inner_map_fds = calloc(max_entries, value_size);
|
/linux-master/tools/lib/bpf/ |
H A D | libbpf_probes.c | 285 int key_size, value_size, max_entries; local 290 value_size = sizeof(__u32); 295 value_size = sizeof(__u64); 299 value_size = sizeof(__u64); 305 value_size = sizeof(__u64); 318 value_size = 8; 328 value_size = 0; 343 value_size = 0; 388 fd = bpf_map_create(map_type, NULL, key_size, value_size, max_entries, &opts);
|
H A D | bpf_gen_internal.h | 60 __u32 key_size, __u32 value_size, __u32 max_entries, 66 void bpf_gen__map_update_elem(struct bpf_gen *gen, int map_idx, void *value, __u32 value_size);
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | bperf_cgroup.bpf.c | 19 __uint(value_size, sizeof(int)); 27 __uint(value_size, sizeof(__u32)); 35 __uint(value_size, sizeof(struct bpf_perf_event_value)); 43 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
H A D | lock_contention.bpf.c | 26 __uint(value_size, sizeof(__u64)); 42 __uint(value_size, sizeof(struct tstamp_data)); 50 __uint(value_size, sizeof(struct contention_data)); 57 __uint(value_size, sizeof(struct contention_task_data)); 64 __uint(value_size, sizeof(__u32)); 71 __uint(value_size, sizeof(__u8)); 78 __uint(value_size, sizeof(__u8)); 85 __uint(value_size, sizeof(__u8)); 92 __uint(value_size, sizeof(__u8)); 99 __uint(value_size, sizeo [all...] |
/linux-master/tools/bpf/bpftool/ |
H A D | map.c | 65 return malloc(round_up(info->value_size, 8) * 68 return malloc(info->value_size); 101 step = round_up(map_info->value_size, 8); 141 print_hex_data_json(value, info->value_size); 159 step = round_up(info->value_size, 8); 173 info->value_size); 253 break_names = info->key_size > 16 || info->value_size > 16; 254 single_line = info->key_size + info->value_size <= 24 && 264 if (info->value_size) { 270 fprint_hex(stdout, value, info->value_size, " "); 340 parse_elem(char **argv, struct bpf_map_info *info, void *key, void *value, __u32 key_size, __u32 value_size, __u32 *flags, __u32 **value_fd) argument 1243 __u32 key_size = 0, value_size = 0, max_entries = 0; local [all...] |
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_htab_mem.c | 40 u32 value_size; member in struct:htab_mem_args 44 .value_size = 8, 68 args.value_size = strtoul(arg, NULL, 10); 69 if (args.value_size > 4096) { 70 fprintf(stderr, "too big value size %u\n", args.value_size); 181 bpf_map__set_value_size(map, args.value_size);
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_dpipe.c | 137 match_value->value_size = sizeof(u32); 138 match_value->value = kmalloc(match_value->value_size, GFP_KERNEL); 143 action_value->value_size = sizeof(u32); 144 action_value->value = kmalloc(action_value->value_size, GFP_KERNEL); 421 match_value->value_size = sizeof(u32); 422 match_value->value = kmalloc(match_value->value_size, GFP_KERNEL); 432 match_value->value_size = sizeof(u32); 435 match_value->value_size = sizeof(struct in6_addr); 442 match_value->value = kmalloc(match_value->value_size, GFP_KERNEL); 447 action_value->value_size [all...] |
/linux-master/tools/hv/ |
H A D | hv_kvp_daemon.c | 322 const __u8 *value, int value_size) 330 (value_size > HV_KVP_EXCHANGE_MAX_VALUE_SIZE)) 349 memcpy(record[i].value, value, value_size); 367 memcpy(record[i].value, value, value_size); 376 int value_size) 383 (value_size > HV_KVP_EXCHANGE_MAX_VALUE_SIZE)) 400 memcpy(value, record[i].value, value_size); 408 __u8 *value, int value_size) 423 memcpy(value, record[index].value, value_size); 1834 hv_msg->body.kvp_set.data.value_size)) 321 kvp_key_add_or_modify(int pool, const __u8 *key, int key_size, const __u8 *value, int value_size) argument 375 kvp_get_value(int pool, const __u8 *key, int key_size, __u8 *value, int value_size) argument 407 kvp_pool_enumerate(int pool, int index, __u8 *key, int key_size, __u8 *value, int value_size) argument [all...] |