Lines Matching defs:smap

24 	struct bpf_local_storage_map *smap;
31 smap = (struct bpf_local_storage_map *)map;
32 return bpf_local_storage_lookup(sk_storage, smap, cacheit_lockit);
135 struct bpf_local_storage_map *smap,
140 copy_selem = bpf_selem_alloc(smap, newsk, NULL, true, GFP_ATOMIC);
144 if (btf_record_has_field(smap->map.record, BPF_SPIN_LOCK))
145 copy_map_value_locked(&smap->map, SDATA(copy_selem)->data,
148 copy_map_value(&smap->map, SDATA(copy_selem)->data,
171 struct bpf_local_storage_map *smap;
174 smap = rcu_dereference(SDATA(selem)->smap);
175 if (!(smap->map.map_flags & BPF_F_CLONE))
183 map = bpf_map_inc_not_zero(&smap->map);
187 copy_selem = bpf_sk_storage_clone_elem(newsk, smap, selem);
195 bpf_selem_link_map(smap, copy_selem);
198 ret = bpf_local_storage_alloc(newsk, smap, copy_selem, GFP_ATOMIC);
200 bpf_selem_free(copy_selem, smap, true);
201 atomic_sub(smap->elem_size,
275 static int bpf_sk_storage_charge(struct bpf_local_storage_map *smap,
292 static void bpf_sk_storage_uncharge(struct bpf_local_storage_map *smap,
548 struct bpf_local_storage_map *smap;
557 smap = rcu_dereference(sdata->smap);
558 if (nla_put_u32(skb, SK_DIAG_BPF_STORAGE_MAP_ID, smap->map.id))
562 smap->map.value_size,
567 if (btf_record_has_field(smap->map.record, BPF_SPIN_LOCK))
568 copy_map_value_locked(&smap->map, nla_data(nla_value),
571 copy_map_value(&smap->map, nla_data(nla_value), sdata->data);
589 struct bpf_local_storage_map *smap;
609 smap = rcu_dereference(SDATA(selem)->smap);
610 diag_size += nla_value_size(smap->map.value_size);
716 struct bpf_local_storage_map *smap;
721 smap = (struct bpf_local_storage_map *)info->map;
722 n_buckets = 1U << smap->bucket_log;
734 b = &smap->buckets[bucket_id++];
747 for (i = bucket_id; i < (1U << smap->bucket_log); i++) {
748 b = &smap->buckets[i];