Searched refs:value_size (Results 51 - 75 of 147) sorted by relevance

123456

/linux-master/tools/perf/util/bpf_skel/
H A Doff_cpu.bpf.c38 __uint(value_size, MAX_STACKS * sizeof(__u64));
52 __uint(value_size, sizeof(__u64));
59 __uint(value_size, sizeof(__u8));
66 __uint(value_size, sizeof(__u8));
73 __uint(value_size, sizeof(__u8));
H A Dkwork_top.bpf.c60 __uint(value_size, sizeof(struct time_data));
67 __uint(value_size, sizeof(struct task_data));
74 __uint(value_size, sizeof(struct work_data));
81 __uint(value_size, sizeof(u8));
H A Dkwork_trace.bpf.c38 __uint(value_size, MAX_KWORKNAME);
45 __uint(value_size, sizeof(__u64));
52 __uint(value_size, sizeof(struct report_data));
59 __uint(value_size, sizeof(__u8));
66 __uint(value_size, MAX_KWORKNAME);
/linux-master/tools/testing/selftests/bpf/progs/
H A Dbloom_filter_bench.c20 /* max entries and value_size will be set programmatically.
27 /* max entries, value_size, and # of hash functions will be set
36 /* max entries, key_size, and value_size, will be set
56 __u8 value_size; variable
126 for (i = 0; i < 1024; i++, index += value_size) {
H A Dtimer_mim_reject.c30 __uint(value_size, sizeof(int));
H A Dtest_unpriv_bpf_disabled.c58 __uint(value_size, sizeof(__u32));
H A Dtailcall_bpf2bpf3.c11 __uint(value_size, sizeof(__u32));
H A Dxsk_xdp_progs.c13 __uint(value_size, sizeof(int));
H A Daccess_map_in_map.c12 __uint(value_size, 4);
H A Dmap_in_map_btf.c30 __uint(value_size, 4);
H A Dmap_ptr_kern.c33 __u32 value_size; member in struct:bpf_map
39 __u32 value_size, __u32 max_entries)
43 VERIFY(map->value_size == value_size);
55 VERIFY(indirect->value_size == direct->value_size);
63 __u32 key_size, __u32 value_size, __u32 max_entries)
66 VERIFY(check_bpf_map_fields(indirect, key_size, value_size,
38 check_bpf_map_fields(struct bpf_map *map, __u32 key_size, __u32 value_size, __u32 max_entries) argument
62 check(struct bpf_map *indirect, struct bpf_map *direct, __u32 key_size, __u32 value_size, __u32 max_entries) argument
H A Dtest_get_stack_rawtp.c22 __uint(value_size, sizeof(__u32));
38 * If we have value_size = MAX_STACK_RAWTP * sizeof(__u64),
H A Dtimer_mim.c29 __uint(value_size, sizeof(int));
/linux-master/drivers/staging/vc04_services/vchiq-mmal/
H A Dmmal-vchiq.h142 u32 value_size);
148 u32 *value_size);
/linux-master/drivers/extcon/
H A Dextcon-rtk-type-c.c900 int value_size = 4; local
901 int value_mask = (BIT(value_size) - 1);
905 cc1_0p2v = get_value((buf[0] >> value_size * 0) & value_mask);
906 cc1_0p8v = get_value((buf[0] >> value_size * 1) & value_mask);
907 cc1_2p6v = get_value((buf[1] >> value_size * 0) & value_mask);
908 cc1_0p66v = get_value((buf[1] >> value_size * 1) & value_mask);
909 cc1_1p23v = get_value((buf[2] >> value_size * 0) & value_mask);
911 cc2_0p2v = get_value((buf[3] >> value_size * 0) & value_mask);
912 cc2_0p8v = get_value((buf[3] >> value_size * 1) & value_mask);
913 cc2_2p6v = get_value((buf[4] >> value_size *
984 int value_size = 0; local
[all...]
/linux-master/drivers/md/persistent-data/
H A Ddm-btree-spine.c38 size_t value_size; local
59 value_size = le32_to_cpu(h->value_size);
62 (sizeof(__le64) + value_size) * max_entries > block_size) {
H A Ddm-btree-remove.c63 uint32_t value_size = le32_to_cpu(n->header.value_size); local
74 (nr_entries - shift) * value_size);
82 nr_entries * value_size); local
89 uint32_t value_size = le32_to_cpu(left->header.value_size); local
91 if (value_size != le32_to_cpu(right->header.value_size)) {
109 shift * value_size); local
121 shift * value_size); local
133 uint32_t value_size = le32_to_cpu(n->header.value_size); local
144 nr_to_copy * value_size); local
[all...]
H A Ddm-btree.c86 static int insert_at(size_t value_size, struct btree_node *node, unsigned int index,
105 array_insert(value_base(node), value_size, nr_entries, index, value); local
117 static uint32_t calc_max_entries(size_t value_size, size_t block_size) argument
120 size_t elt_size = sizeof(uint64_t) + value_size; /* key + value */
149 n->header.value_size = cpu_to_le32(info->value_type.size);
345 uint64_t *result_key, void *v, size_t value_size)
369 memcpy(v, value_ptr(ro_node(s), i), value_size);
511 size_t value_size = le32_to_cpu(dest->header.value_size); local
514 memcpy(value_ptr(dest, dest_offset), value_ptr(src, src_offset), count * value_size); local
343 btree_lookup_raw(struct ro_spine *s, dm_block_t block, uint64_t key, int (*search_fn)(struct btree_node *, uint64_t), uint64_t *result_key, void *v, size_t value_size) argument
525 size_t value_size = le32_to_cpu(dest->header.value_size); local
528 memmove(value_ptr(dest, dest_offset), value_ptr(src, src_offset), count * value_size); local
[all...]
/linux-master/kernel/bpf/
H A Dreuseport_array.c43 if (attr->value_size != sizeof(u32) &&
44 attr->value_size != sizeof(u64))
171 if (map->value_size != sizeof(u64))
249 if (map->value_size == sizeof(u64)) {
H A Dlocal_storage.c167 new = bpf_map_kmalloc_node(map, struct_size(new, data, map->value_size),
173 memcpy(&new->data[0], value, map->value_size);
198 * access 'value_size' of them, so copying rounded areas
201 size = round_up(_map->value_size, 8);
229 /* the user space will provide round_up(value_size, 8) bytes that
231 * value_size of it. During lookup the same extra bytes will be
235 size = round_up(_map->value_size, 8);
302 if (attr->value_size == 0)
305 if (attr->value_size > max_value_size)
487 size = sizeof(struct bpf_storage_buffer) + map->value_size;
[all...]
/linux-master/samples/bpf/
H A Dtracex6.bpf.c10 __uint(value_size, sizeof(u32));
H A Dxdp_router_ipv4.bpf.c45 __uint(value_size, sizeof(struct trie_value));
69 __uint(value_size, sizeof(int));
H A Dtrace_event_kern.c30 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
/linux-master/arch/sparc/include/asm/
H A Doplib_32.h162 int value_size);
/linux-master/drivers/net/ethernet/netronome/nfp/bpf/
H A Dfw.h81 __be32 value_size; /* in bytes */ member in struct:cmsg_req_map_alloc_tbl

Completed in 272 milliseconds

123456