/linux-master/tools/hv/ |
H A D | hv_kvp_daemon.c | 322 const __u8 *value, int value_size) 330 (value_size > HV_KVP_EXCHANGE_MAX_VALUE_SIZE)) 349 memcpy(record[i].value, value, value_size); 367 memcpy(record[i].value, value, value_size); 376 int value_size) 383 (value_size > HV_KVP_EXCHANGE_MAX_VALUE_SIZE)) 400 memcpy(value, record[i].value, value_size); 408 __u8 *value, int value_size) 423 memcpy(value, record[index].value, value_size); 1834 hv_msg->body.kvp_set.data.value_size)) 321 kvp_key_add_or_modify(int pool, const __u8 *key, int key_size, const __u8 *value, int value_size) argument 375 kvp_get_value(int pool, const __u8 *key, int key_size, __u8 *value, int value_size) argument 407 kvp_pool_enumerate(int pool, int index, __u8 *key, int key_size, __u8 *value, int value_size) argument [all...] |
/linux-master/tools/include/uapi/linux/ |
H A D | bpf.h | 652 * *value_size* * *count*. 703 * *value_size* * *count*. 1458 __u32 value_size; /* size of value in bytes */ member in struct:bpf_attr::__anon3130 6582 __u32 value_size; member in struct:bpf_map_info
|
/linux-master/tools/lib/bpf/ |
H A D | bpf.c | 171 __u32 value_size, 190 attr.value_size = value_size; 168 bpf_map_create(enum bpf_map_type map_type, const char *map_name, __u32 key_size, __u32 value_size, __u32 max_entries, const struct bpf_map_create_opts *opts) argument
|
H A D | bpf.h | 64 __u32 value_size, 205 * *value_size* * *count*.
|
H A D | bpf_gen_internal.h | 60 __u32 key_size, __u32 value_size, __u32 max_entries, 66 void bpf_gen__map_update_elem(struct bpf_gen *gen, int map_idx, void *value, __u32 value_size);
|
H A D | gen_loader.c | 451 __u32 key_size, __u32 value_size, __u32 max_entries, 462 attr.value_size = value_size; 500 debug_ret(gen, "map_create %s idx %d type %d value_size %d value_btf_id %d", 501 attr.map_name, map_idx, map_type, value_size, 1035 __u32 value_size) 1045 value = add_data(gen, pvalue, value_size); 1050 * bpf_probe_read_kernel(value, value_size, initial_value); 1052 * bpf_copy_from_user(value, value_size, initial_value); 1062 emit(gen, BPF_MOV64_IMM(BPF_REG_2, value_size)); 448 bpf_gen__map_create(struct bpf_gen *gen, enum bpf_map_type map_type, const char *map_name, __u32 key_size, __u32 value_size, __u32 max_entries, struct bpf_map_create_opts *map_attr, int map_idx) argument 1034 bpf_gen__map_update_elem(struct bpf_gen *gen, int map_idx, void *pvalue, __u32 value_size) argument [all...] |
H A D | libbpf.c | 514 unsigned int value_size; member in struct:bpf_map_def 1113 map->def.value_size = kern_vtype->size; 1344 map->def.value_size = type->size; 1716 return array_map_mmap_sz(map->def.value_size, map->def.max_entries); 1863 def->value_size = data_sz; 2514 } else if (strcmp(name, "value_size") == 0) { 2519 if (map_def->value_size && map_def->value_size != sz) { 2521 map_name, map_def->value_size, sz); 2524 map_def->value_size [all...] |
H A D | libbpf_internal.h | 270 __u32 value_size; member in struct:btf_map_def
|
H A D | libbpf_probes.c | 285 int key_size, value_size, max_entries; local 290 value_size = sizeof(__u32); 295 value_size = sizeof(__u64); 299 value_size = sizeof(__u64); 305 value_size = sizeof(__u64); 318 value_size = 8; 328 value_size = 0; 343 value_size = 0; 388 fd = bpf_map_create(map_type, NULL, key_size, value_size, max_entries, &opts);
|
H A D | linker.c | 1544 if (main_def->value_size != extra_def->value_size) { 1545 reason = "value_size";
|
H A D | skel_internal.h | 222 __u32 value_size, 233 attr.value_size = value_size; 219 skel_map_create(enum bpf_map_type map_type, const char *map_name, __u32 key_size, __u32 value_size, __u32 max_entries) argument
|
/linux-master/tools/perf/util/ |
H A D | bpf_counter.c | 315 (map_info.value_size == sizeof(struct perf_event_attr_map_entry));
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | bperf_cgroup.bpf.c | 19 __uint(value_size, sizeof(int)); 27 __uint(value_size, sizeof(__u32)); 35 __uint(value_size, sizeof(struct bpf_perf_event_value)); 43 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
H A D | bperf_follower.bpf.c | 11 __uint(value_size, sizeof(struct bpf_perf_event_value)); 18 __uint(value_size, sizeof(struct bpf_perf_event_value)); 25 __uint(value_size, sizeof(__u32));
|
H A D | bperf_leader.bpf.c | 10 __uint(value_size, sizeof(int)); 17 __uint(value_size, sizeof(struct bpf_perf_event_value)); 24 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
H A D | bpf_prog_profiler.bpf.c | 11 __uint(value_size, sizeof(int)); 18 __uint(value_size, sizeof(struct bpf_perf_event_value)); 26 __uint(value_size, sizeof(struct bpf_perf_event_value));
|
H A D | func_latency.bpf.c | 13 __uint(value_size, sizeof(__u64)); 20 __uint(value_size, sizeof(__u8)); 27 __uint(value_size, sizeof(__u8)); 34 __uint(value_size, sizeof(__u64));
|
H A D | kwork_top.bpf.c | 60 __uint(value_size, sizeof(struct time_data)); 67 __uint(value_size, sizeof(struct task_data)); 74 __uint(value_size, sizeof(struct work_data)); 81 __uint(value_size, sizeof(u8));
|
H A D | kwork_trace.bpf.c | 38 __uint(value_size, MAX_KWORKNAME); 45 __uint(value_size, sizeof(__u64)); 52 __uint(value_size, sizeof(struct report_data)); 59 __uint(value_size, sizeof(__u8)); 66 __uint(value_size, MAX_KWORKNAME);
|
H A D | lock_contention.bpf.c | 26 __uint(value_size, sizeof(__u64)); 42 __uint(value_size, sizeof(struct tstamp_data)); 50 __uint(value_size, sizeof(struct contention_data)); 57 __uint(value_size, sizeof(struct contention_task_data)); 64 __uint(value_size, sizeof(__u32)); 71 __uint(value_size, sizeof(__u8)); 78 __uint(value_size, sizeof(__u8)); 85 __uint(value_size, sizeof(__u8)); 92 __uint(value_size, sizeof(__u8)); 99 __uint(value_size, sizeo [all...] |
H A D | off_cpu.bpf.c | 38 __uint(value_size, MAX_STACKS * sizeof(__u64)); 52 __uint(value_size, sizeof(__u64)); 59 __uint(value_size, sizeof(__u8)); 66 __uint(value_size, sizeof(__u8)); 73 __uint(value_size, sizeof(__u8));
|
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bloom_filter_map.c | 41 __u8 value_size; member in struct:__anon5763 45 .value_size = 8, 59 { "value_size", ARG_VALUE_SIZE, "VALUE_SIZE", 0, 93 args.value_size = ret; 136 val_size = args.value_size; 243 if (args.value_size < 8) { 244 __u64 nr_unique_entries = 1ULL << (args.value_size * 8); 279 bpf_map__set_value_size(skel->maps.array_map, args.value_size); 281 bpf_map__set_value_size(skel->maps.bloom_map, args.value_size); 283 bpf_map__set_value_size(skel->maps.hashmap, args.value_size); [all...] |
H A D | bench_htab_mem.c | 40 u32 value_size; member in struct:htab_mem_args 44 .value_size = 8, 68 args.value_size = strtoul(arg, NULL, 10); 69 if (args.value_size > 4096) { 70 fprintf(stderr, "too big value size %u\n", args.value_size); 181 bpf_map__set_value_size(map, args.value_size);
|
/linux-master/tools/testing/selftests/bpf/map_tests/ |
H A D | array_map_batch_ops.c | 76 int err, step, value_size; local 88 value_size = sizeof(__s64); 90 value_size *= nr_cpus; 93 values = calloc(max_entries, value_size); 104 memset(values, 0, max_entries * value_size); 115 values + total * value_size,
|
H A D | htab_map_batch_ops.c | 84 int err, step, value_size; local 97 value_size = is_pcpu ? sizeof(value) : sizeof(int); 125 memset(values, 0, max_entries * value_size); 144 memset(values, 0, max_entries * value_size); 156 total * value_size, 210 memset(values, 0, max_entries * value_size); 219 total * value_size,
|