• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/net/iucv/

Lines Matching refs:iucv_sk

107 		if (!memcmp(&iucv_sk(sk)->src_name, nm, 8))
149 struct iucv_sock *iucv = iucv_sk(sk);
221 INIT_LIST_HEAD(&iucv_sk(sk)->accept_q);
222 skb_queue_head_init(&iucv_sk(sk)->send_skb_q);
223 skb_queue_head_init(&iucv_sk(sk)->backlog_skb_q);
224 iucv_sk(sk)->send_tag = 0;
278 list_add_tail(&iucv_sk(sk)->accept_q, &iucv_sk(parent)->accept_q);
279 iucv_sk(sk)->parent = parent;
285 list_del_init(&iucv_sk(sk)->accept_q);
286 iucv_sk(sk)->parent->sk_ack_backlog--;
287 iucv_sk(sk)->parent = NULL;
296 list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) {
379 iucv = iucv_sk(sk);
408 struct iucv_sock *iucv = iucv_sk(sk);
454 iucv = iucv_sk(sk);
465 memcpy(iucv_sk(sk)->dst_user_id, sa->siucv_user_id, 8);
466 memcpy(iucv_sk(sk)->dst_name, sa->siucv_name, 8);
469 low_nmcpy(user_data, iucv_sk(sk)->src_name);
472 iucv = iucv_sk(sk);
586 memcpy(siucv->siucv_user_id, iucv_sk(sk)->dst_user_id, 8);
587 memcpy(siucv->siucv_name, &iucv_sk(sk)->dst_name, 8);
589 memcpy(siucv->siucv_user_id, iucv_sk(sk)->src_user_id, 8);
590 memcpy(siucv->siucv_name, iucv_sk(sk)->src_name, 8);
603 struct iucv_sock *iucv = iucv_sk(sk);
667 struct iucv_sock *iucv = iucv_sk(sk);
713 rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q);
716 skb_queue_head(&iucv_sk(sk)->backlog_skb_q,
720 rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q);
735 list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) {
786 struct iucv_sock *iucv = iucv_sk(sk);
828 err = iucv_path_quiesce(iucv_sk(sk)->path, NULL);
854 if (iucv_sk(sk)->path) {
855 iucv_path_sever(iucv_sk(sk)->path, NULL);
856 iucv_path_free(iucv_sk(sk)->path);
857 iucv_sk(sk)->path = NULL;
884 !memcmp(&iucv_sk(sk)->src_name, src_name, 8)) {
889 iucv = iucv_sk(sk);
921 niucv = iucv_sk(nsk);
997 struct iucv_sock *iucv = iucv_sk(sk);
1044 skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, fskb);
1051 skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb);
1053 skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb);
1063 struct sk_buff_head *list = &iucv_sk(sk)->send_skb_q;
1077 skb_unlink(this, &iucv_sk(sk)->send_skb_q);
1082 if (skb_queue_empty(&iucv_sk(sk)->send_skb_q)) {
1094 if (!list_empty(&iucv_sk(sk)->accept_q))