/linux-master/tools/perf/util/bpf_skel/ |
H A D | off_cpu.bpf.c | 38 __uint(value_size, MAX_STACKS * sizeof(__u64)); 52 __uint(value_size, sizeof(__u64)); 59 __uint(value_size, sizeof(__u8)); 66 __uint(value_size, sizeof(__u8)); 73 __uint(value_size, sizeof(__u8));
|
H A D | kwork_top.bpf.c | 60 __uint(value_size, sizeof(struct time_data)); 67 __uint(value_size, sizeof(struct task_data)); 74 __uint(value_size, sizeof(struct work_data)); 81 __uint(value_size, sizeof(u8));
|
H A D | kwork_trace.bpf.c | 38 __uint(value_size, MAX_KWORKNAME); 45 __uint(value_size, sizeof(__u64)); 52 __uint(value_size, sizeof(struct report_data)); 59 __uint(value_size, sizeof(__u8)); 66 __uint(value_size, MAX_KWORKNAME);
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | bloom_filter_bench.c | 20 /* max entries and value_size will be set programmatically. 27 /* max entries, value_size, and # of hash functions will be set 36 /* max entries, key_size, and value_size, will be set 56 __u8 value_size; variable 126 for (i = 0; i < 1024; i++, index += value_size) {
|
H A D | timer_mim_reject.c | 30 __uint(value_size, sizeof(int));
|
H A D | test_unpriv_bpf_disabled.c | 58 __uint(value_size, sizeof(__u32));
|
H A D | tailcall_bpf2bpf3.c | 11 __uint(value_size, sizeof(__u32));
|
H A D | xsk_xdp_progs.c | 13 __uint(value_size, sizeof(int));
|
H A D | access_map_in_map.c | 12 __uint(value_size, 4);
|
H A D | map_in_map_btf.c | 30 __uint(value_size, 4);
|
H A D | map_ptr_kern.c | 33 __u32 value_size; member in struct:bpf_map 39 __u32 value_size, __u32 max_entries) 43 VERIFY(map->value_size == value_size); 55 VERIFY(indirect->value_size == direct->value_size); 63 __u32 key_size, __u32 value_size, __u32 max_entries) 66 VERIFY(check_bpf_map_fields(indirect, key_size, value_size, 38 check_bpf_map_fields(struct bpf_map *map, __u32 key_size, __u32 value_size, __u32 max_entries) argument 62 check(struct bpf_map *indirect, struct bpf_map *direct, __u32 key_size, __u32 value_size, __u32 max_entries) argument
|
H A D | test_get_stack_rawtp.c | 22 __uint(value_size, sizeof(__u32)); 38 * If we have value_size = MAX_STACK_RAWTP * sizeof(__u64),
|
H A D | timer_mim.c | 29 __uint(value_size, sizeof(int));
|
/linux-master/drivers/staging/vc04_services/vchiq-mmal/ |
H A D | mmal-vchiq.h | 142 u32 value_size); 148 u32 *value_size);
|
/linux-master/drivers/extcon/ |
H A D | extcon-rtk-type-c.c | 900 int value_size = 4; local 901 int value_mask = (BIT(value_size) - 1); 905 cc1_0p2v = get_value((buf[0] >> value_size * 0) & value_mask); 906 cc1_0p8v = get_value((buf[0] >> value_size * 1) & value_mask); 907 cc1_2p6v = get_value((buf[1] >> value_size * 0) & value_mask); 908 cc1_0p66v = get_value((buf[1] >> value_size * 1) & value_mask); 909 cc1_1p23v = get_value((buf[2] >> value_size * 0) & value_mask); 911 cc2_0p2v = get_value((buf[3] >> value_size * 0) & value_mask); 912 cc2_0p8v = get_value((buf[3] >> value_size * 1) & value_mask); 913 cc2_2p6v = get_value((buf[4] >> value_size * 984 int value_size = 0; local [all...] |
/linux-master/drivers/md/persistent-data/ |
H A D | dm-btree-spine.c | 38 size_t value_size; local 59 value_size = le32_to_cpu(h->value_size); 62 (sizeof(__le64) + value_size) * max_entries > block_size) {
|
H A D | dm-btree-remove.c | 63 uint32_t value_size = le32_to_cpu(n->header.value_size); local 74 (nr_entries - shift) * value_size); 82 nr_entries * value_size); local 89 uint32_t value_size = le32_to_cpu(left->header.value_size); local 91 if (value_size != le32_to_cpu(right->header.value_size)) { 109 shift * value_size); local 121 shift * value_size); local 133 uint32_t value_size = le32_to_cpu(n->header.value_size); local 144 nr_to_copy * value_size); local [all...] |
H A D | dm-btree.c | 86 static int insert_at(size_t value_size, struct btree_node *node, unsigned int index, 105 array_insert(value_base(node), value_size, nr_entries, index, value); local 117 static uint32_t calc_max_entries(size_t value_size, size_t block_size) argument 120 size_t elt_size = sizeof(uint64_t) + value_size; /* key + value */ 149 n->header.value_size = cpu_to_le32(info->value_type.size); 345 uint64_t *result_key, void *v, size_t value_size) 369 memcpy(v, value_ptr(ro_node(s), i), value_size); 511 size_t value_size = le32_to_cpu(dest->header.value_size); local 514 memcpy(value_ptr(dest, dest_offset), value_ptr(src, src_offset), count * value_size); local 343 btree_lookup_raw(struct ro_spine *s, dm_block_t block, uint64_t key, int (*search_fn)(struct btree_node *, uint64_t), uint64_t *result_key, void *v, size_t value_size) argument 525 size_t value_size = le32_to_cpu(dest->header.value_size); local 528 memmove(value_ptr(dest, dest_offset), value_ptr(src, src_offset), count * value_size); local [all...] |
/linux-master/kernel/bpf/ |
H A D | reuseport_array.c | 43 if (attr->value_size != sizeof(u32) && 44 attr->value_size != sizeof(u64)) 171 if (map->value_size != sizeof(u64)) 249 if (map->value_size == sizeof(u64)) {
|
H A D | local_storage.c | 167 new = bpf_map_kmalloc_node(map, struct_size(new, data, map->value_size), 173 memcpy(&new->data[0], value, map->value_size); 198 * access 'value_size' of them, so copying rounded areas 201 size = round_up(_map->value_size, 8); 229 /* the user space will provide round_up(value_size, 8) bytes that 231 * value_size of it. During lookup the same extra bytes will be 235 size = round_up(_map->value_size, 8); 302 if (attr->value_size == 0) 305 if (attr->value_size > max_value_size) 487 size = sizeof(struct bpf_storage_buffer) + map->value_size; [all...] |
/linux-master/samples/bpf/ |
H A D | tracex6.bpf.c | 10 __uint(value_size, sizeof(u32));
|
H A D | xdp_router_ipv4.bpf.c | 45 __uint(value_size, sizeof(struct trie_value)); 69 __uint(value_size, sizeof(int));
|
H A D | trace_event_kern.c | 30 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
/linux-master/arch/sparc/include/asm/ |
H A D | oplib_32.h | 162 int value_size);
|
/linux-master/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | fw.h | 81 __be32 value_size; /* in bytes */ member in struct:cmsg_req_map_alloc_tbl
|