/linux-master/fs/overlayfs/ |
H A D | copy_up.c | 79 ssize_t list_size, size, value_size = 0; local 134 size = ovl_do_getxattr(oldpath, name, value, value_size); 143 if (size > value_size) { 153 value_size = size;
|
/linux-master/fs/xfs/scrub/ |
H A D | attr.c | 71 size_t value_size) 99 if (ab->value_sz >= value_size) 108 new_val = kvmalloc(value_size, XCHK_GFP_FLAGS); 113 ab->value_sz = value_size; 69 xchk_setup_xattr_buf( struct xfs_scrub *sc, size_t value_size) argument
|
/linux-master/include/linux/ |
H A D | bpf.h | 265 u32 value_size; member in struct:bpf_map 492 bpf_obj_memcpy(map->record, dst, src, map->value_size, false); 497 bpf_obj_memcpy(map->record, dst, src, map->value_size, true); 522 bpf_obj_memzero(map->record, dst, map->value_size);
|
H A D | btf.h | 221 u32 field_mask, u32 value_size);
|
/linux-master/include/net/ |
H A D | devlink.h | 276 * @value_size: value size 287 unsigned int value_size; member in struct:devlink_dpipe_value
|
/linux-master/include/uapi/linux/ |
H A D | bpf.h | 652 * *value_size* * *count*. 703 * *value_size* * *count*. 1458 __u32 value_size; /* size of value in bytes */ member in struct:bpf_attr::__anon2589 6582 __u32 value_size; member in struct:bpf_map_info
|
H A D | hyperv.h | 359 __u32 value_size; member in struct:hv_kvp_exchg_msg_value
|
/linux-master/kernel/bpf/ |
H A D | arena.c | 101 if (attr->key_size || attr->value_size || attr->max_entries == 0 ||
|
H A D | arraymap.c | 59 attr->value_size == 0 || 73 /* avoid overflow on round_up(map->value_size) */ 74 if (attr->value_size > INT_MAX) 89 elem_size = round_up(attr->value_size, 8); 181 if (off >= map->value_size) 275 * access 'value_size' of them, so copying rounded areas 370 /* the user space will provide round_up(value_size, 8) bytes that 372 * value_size of it. During lookup the same extra bytes will be 796 if (attr->value_size != sizeof(u32))
|
H A D | bloom_filter.c | 24 u32 value_size, u32 index) 28 if (likely(value_size % 4 == 0)) 29 h = jhash2(value, value_size / 4, bloom->hash_seed + index); 31 h = jhash(value, value_size, bloom->hash_seed + index); 43 h = hash(bloom, value, map->value_size, i); 61 h = hash(bloom, value, map->value_size, i); 86 if (attr->value_size > KMALLOC_MAX_SIZE) 87 /* if value_size is bigger, the user space won't be able to 101 if (attr->key_size != 0 || attr->value_size == 0 || 23 hash(struct bpf_bloom_filter *bloom, void *value, u32 value_size, u32 index) argument
|
H A D | bpf_local_storage.c | 95 memset(SDATA(selem)->data, 0, smap->map.value_size); 676 attr->key_size != sizeof(int) || !attr->value_size || 681 if (attr->value_size > BPF_LOCAL_STORAGE_MAX_VALUE_SIZE) 798 sdata.data[attr->value_size]);
|
H A D | bpf_struct_ops.c | 451 memset(value, 0, map->value_size); 459 memcpy(uvalue, st_map->uvalue, map->value_size); 635 memcpy(uvalue, value, map->value_size); 784 memset(uvalue, 0, map->value_size); 785 memset(kvalue, 0, map->value_size); 827 value = kmalloc(map->value_size, GFP_USER | __GFP_NOWARN); 937 if (attr->value_size != vt->size) {
|
H A D | btf.c | 3796 u32 field_mask, u32 value_size) 3822 if (info_arr[i].off + field_type_size > value_size) { 3823 WARN_ONCE(1, "verifier bug off %d size %d", info_arr[i].off, value_size); 3795 btf_parse_fields(const struct btf *btf, const struct btf_type *t, u32 field_mask, u32 value_size) argument
|
H A D | cpumap.c | 86 u32 value_size = attr->value_size; local 91 (value_size != offsetofend(struct bpf_cpumap_val, qsize) && 92 value_size != offsetofend(struct bpf_cpumap_val, bpf_prog.fd)) || 537 memcpy(&cpumap_value, value, map->value_size);
|
H A D | devmap.c | 112 u32 valsize = attr->value_size; 921 /* already verified value_size <= sizeof val */ 922 memcpy(&val, value, map->value_size); 965 /* already verified value_size <= sizeof val */ 966 memcpy(&val, value, map->value_size);
|
H A D | hashtab.c | 335 u32 size = round_up(htab->map.value_size, 8); 448 * value_size == 0 may be allowed in the future to use map as a set 451 attr->value_size == 0) 454 if ((u64)attr->key_size + attr->value_size >= KMALLOC_MAX_SIZE - 456 /* if key_size + value_size is bigger, the user space won't be 516 htab->elem_size += round_up(htab->map.value_size, 8); 591 round_up(htab->map.value_size, 8), true); 956 /* copy true value_size bytes */ 959 u32 size = round_up(htab->map.value_size, 8); 1003 u32 size = htab->map.value_size; 1687 u32 bucket_cnt, total, key_size, value_size, roundup_key_size; local 2220 u32 value_size = round_up(htab->map.value_size, 8); local [all...] |
H A D | local_storage.c | 167 new = bpf_map_kmalloc_node(map, struct_size(new, data, map->value_size), 173 memcpy(&new->data[0], value, map->value_size); 198 * access 'value_size' of them, so copying rounded areas 201 size = round_up(_map->value_size, 8); 229 /* the user space will provide round_up(value_size, 8) bytes that 231 * value_size of it. During lookup the same extra bytes will be 235 size = round_up(_map->value_size, 8); 302 if (attr->value_size == 0) 305 if (attr->value_size > max_value_size) 487 size = sizeof(struct bpf_storage_buffer) + map->value_size; [all...] |
H A D | log.c | 721 reg->map_ptr->value_size);
|
H A D | lpm_trie.c | 289 size += trie->map.value_size; 300 trie->map.value_size); 557 attr->value_size < LPM_VAL_SIZE_MIN || 558 attr->value_size > LPM_VAL_SIZE_MAX) 729 trie->map.value_size;
|
H A D | map_in_map.c | 46 inner_map_meta->value_size = inner_map->value_size; 103 meta0->value_size == meta1->value_size &&
|
H A D | map_iter.c | 104 u32 key_acc_size, value_acc_size, key_size, value_size; local 129 value_size = map->value_size; 131 value_size = round_up(map->value_size, 8) * num_possible_cpus(); 133 if (key_acc_size > key_size || value_acc_size > value_size) {
|
H A D | queue_stack_maps.c | 50 attr->value_size == 0 || 55 if (attr->value_size > KMALLOC_MAX_SIZE) 56 /* if value_size is bigger, the user space won't be able to 71 queue_size = sizeof(*qs) + size * attr->value_size; 109 memset(value, 0, qs->map.value_size); 114 ptr = &qs->elements[qs->tail * qs->map.value_size]; 115 memcpy(value, ptr, qs->map.value_size); 144 memset(value, 0, qs->map.value_size); 153 ptr = &qs->elements[index * qs->map.value_size]; 154 memcpy(value, ptr, qs->map.value_size); [all...] |
H A D | reuseport_array.c | 43 if (attr->value_size != sizeof(u32) && 44 attr->value_size != sizeof(u64)) 171 if (map->value_size != sizeof(u64)) 249 if (map->value_size == sizeof(u64)) {
|
H A D | ringbuf.c | 193 if (attr->key_size || attr->value_size ||
|
H A D | stackmap.c | 48 (u64)smap->map.value_size; 72 u32 value_size = attr->value_size; local 82 value_size < 8 || value_size % 8) 87 if (value_size % sizeof(struct bpf_stack_build_id) || 88 value_size / sizeof(struct bpf_stack_build_id) 91 } else if (value_size / 8 > sysctl_perf_event_max_stack) 286 u32 max_depth = map->value_size / stack_map_data_size(map); 593 memset(value + trace_len, 0, map->value_size 669 u64 value_size = map->value_size; local [all...] |