Lines Matching defs:val
668 int sk_set_peek_off(struct sock *sk, int val);
679 static inline void sk_peek_offset_bwd(struct sock *sk, int val)
684 off = max_t(s32, off - val, 0);
689 static inline void sk_peek_offset_fwd(struct sock *sk, int val)
691 sk_peek_offset_bwd(sk, -val);
1036 static inline void sk_wmem_queued_add(struct sock *sk, int val)
1038 WRITE_ONCE(sk->sk_wmem_queued, sk->sk_wmem_queued + val);
1041 static inline void sk_forward_alloc_add(struct sock *sk, int val)
1044 WRITE_ONCE(sk->sk_forward_alloc, sk->sk_forward_alloc + val);
1411 int val[PROTO_INUSE_NR];
1415 const struct proto *prot, int val)
1417 this_cpu_add(net->core.prot_inuse->val[prot->inuse_idx], val);
1420 static inline void sock_inuse_add(const struct net *net, int val)
1422 this_cpu_add(net->core.prot_inuse->all, val);
1429 const struct proto *prot, int val)
1433 static inline void sock_inuse_add(const struct net *net, int val)
1921 int val = READ_ONCE(sk->sk_tx_queue_mapping);
1923 if (val != NO_QUEUE_MAPPING)
1924 return val;
2458 u32 val;
2463 val = min(sk->sk_sndbuf, sk->sk_wmem_queued >> 1);
2464 val = max_t(u32, val, sk_unused_reserved_mem(sk));
2466 WRITE_ONCE(sk->sk_sndbuf, max_t(u32, val, SOCK_MIN_SNDBUF));
2834 static inline void sk_pacing_shift_update(struct sock *sk, int val)
2836 if (!sk || !sk_fullsock(sk) || READ_ONCE(sk->sk_pacing_shift) == val)
2838 WRITE_ONCE(sk->sk_pacing_shift, val);
2872 void sock_set_rcvbuf(struct sock *sk, int val);
2873 void sock_set_mark(struct sock *sk, u32 val);