Lines Matching defs:nsk
460 static void smc_adjust_sock_bufsizes(struct sock *nsk, struct sock *osk,
463 struct net *nnet = sock_net(nsk);
465 nsk->sk_userlocks = osk->sk_userlocks;
467 nsk->sk_sndbuf = osk->sk_sndbuf;
470 WRITE_ONCE(nsk->sk_sndbuf,
473 WRITE_ONCE(nsk->sk_sndbuf,
477 nsk->sk_rcvbuf = osk->sk_rcvbuf;
480 WRITE_ONCE(nsk->sk_rcvbuf,
483 WRITE_ONCE(nsk->sk_rcvbuf,
488 static void smc_copy_sock_settings(struct sock *nsk, struct sock *osk,
492 nsk->sk_type = osk->sk_type;
493 nsk->sk_sndtimeo = osk->sk_sndtimeo;
494 nsk->sk_rcvtimeo = osk->sk_rcvtimeo;
495 nsk->sk_mark = READ_ONCE(osk->sk_mark);
496 nsk->sk_priority = READ_ONCE(osk->sk_priority);
497 nsk->sk_rcvlowat = osk->sk_rcvlowat;
498 nsk->sk_bound_dev_if = osk->sk_bound_dev_if;
499 nsk->sk_err = osk->sk_err;
501 nsk->sk_flags &= ~mask;
502 nsk->sk_flags |= osk->sk_flags & mask;
504 smc_adjust_sock_bufsizes(nsk, osk, mask);
2679 struct sock *sk = sock->sk, *nsk;
2698 while (!(nsk = smc_accept_dequeue(sk, new_sock))) {
2718 rc = sock_error(nsk);
2727 if (smc_sk(nsk)->use_fallback) {
2728 struct sock *clcsk = smc_sk(nsk)->clcsock->sk;
2734 } else if (!atomic_read(&smc_sk(nsk)->conn.bytes_to_rcv)) {
2735 lock_sock(nsk);
2736 smc_rx_wait(smc_sk(nsk), &timeo, smc_rx_data_available);
2737 release_sock(nsk);