/asus-wl-520gu-7.0.1.45/src/linux/linux/net/netrom/ |
H A D | nr_loopback.c | 45 struct sk_buff *skbn; local 47 if ((skbn = alloc_skb(skb->len, GFP_ATOMIC)) != NULL) { 48 memcpy(skb_put(skbn, skb->len), skb->data, skb->len); 49 skbn->h.raw = skbn->data; 51 skb_queue_tail(&loopback_queue, skbn);
|
H A D | nr_out.c | 47 struct sk_buff *skbn; local 59 if ((skbn = sock_alloc_send_skb(sk, frontlen + NR_MAX_PACKET_SIZE, 0, &err)) == NULL) 62 skb_reserve(skbn, frontlen); 67 memcpy(skb_put(skbn, len), skb->data, len); 71 skb_push(skbn, NR_TRANSPORT_LEN); 72 memcpy(skbn->data, transport, NR_TRANSPORT_LEN); 75 skbn->data[4] |= NR_MORE_FLAG; 77 skb_queue_tail(&sk->write_queue, skbn); /* Throw it on the queue */ 110 struct sk_buff *skb, *skbn; local 115 if ((skbn 134 struct sk_buff *skb, *skbn; local [all...] |
H A D | nr_in.c | 52 struct sk_buff *skbo, *skbn = skb; local 68 if ((skbn = alloc_skb(sk->protinfo.nr->fraglen, GFP_ATOMIC)) == NULL) 71 skbn->h.raw = skbn->data; 74 memcpy(skb_put(skbn, skbo->len), skbo->data, skbo->len); 81 return sock_queue_rcv_skb(sk, skbn); 156 struct sk_buff *skbn; local 240 while ((skbn = skb_dequeue(&sk->protinfo.nr->reseq_queue)) != NULL) { 241 ns = skbn->data[17]; 243 if (nr_queue_rx_frame(sk, skbn, frametyp [all...] |
H A D | nr_dev.c | 90 struct sk_buff *skbn; local 107 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { 113 skb_set_owner_w(skbn, skb->sk); 117 len = skbn->len; 119 if (!nr_route_frame(skbn, NULL)) { 120 kfree_skb(skbn);
|
H A D | nr_subr.c | 223 struct sk_buff *skbn; local 229 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL) 232 skb_reserve(skbn, AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN); 234 dptr = skb_put(skbn, NR_NETWORK_LEN + NR_TRANSPORT_LEN); 265 if (!nr_route_frame(skbn, NULL)) 266 kfree_skb(skbn);
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/net/ax25/ |
H A D | ax25_out.c | 138 struct sk_buff *skbn; local 161 if ((skbn = alloc_skb(paclen + 2 + frontlen, GFP_ATOMIC)) == NULL) { 168 skb_set_owner_w(skbn, skb->sk); 175 skb_reserve(skbn, frontlen + 2); 176 skbn->nh.raw = skbn->data + (skb->nh.raw - skb->data); 177 memcpy(skb_put(skbn, len), skb->data, len); 178 p = skb_push(skbn, 2); 188 skb_reserve(skbn, frontlen + 1); 189 skbn 258 struct sk_buff *skb, *skbn; local 340 struct sk_buff *skbn; local [all...] |
H A D | ax25_in.c | 73 struct sk_buff *skbn, *skbo; local 86 skbn = alloc_skb(AX25_MAX_HEADER_LEN + 89 if (!skbn) { 94 skb_reserve(skbn, AX25_MAX_HEADER_LEN); 96 skbn->dev = ax25->ax25_dev->dev; 97 skbn->h.raw = skbn->data; 98 skbn->nh.raw = skbn->data; 102 memcpy(skb_put(skbn, skb 152 struct sk_buff *skbn = skb_copy(skb, GFP_ATOMIC); local [all...] |
H A D | ax25_route.c | 409 struct sk_buff *skbn; local 416 if ((skbn = skb_realloc_headroom(skb, len)) == NULL) { 422 skb_set_owner_w(skbn, skb->sk); 426 skb = skbn;
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/net/x25/ |
H A D | x25_out.c | 66 struct sk_buff *skbn; local 82 if ((skbn = sock_alloc_send_skb(sk, frontlen + max_len, noblock, &err)) == NULL){ 91 skb_reserve(skbn, frontlen); 96 memcpy(skb_put(skbn, len), skb->data, len); 100 skb_push(skbn, header_len); 101 memcpy(skbn->data, header, header_len); 105 skbn->data[3] |= X25_EXT_M_BIT; 107 skbn->data[2] |= X25_STD_M_BIT; 110 skb_queue_tail(&sk->write_queue, skbn); 146 struct sk_buff *skb, *skbn; local [all...] |
H A D | x25_in.c | 49 struct sk_buff *skbo, *skbn = skb; local 61 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL){ 68 skbn->h.raw = skbn->data; 71 memcpy(skb_put(skbn, skbo->len), skbo->data, skbo->len); 76 memcpy(skb_put(skbn, skbo->len), skbo->data, skbo->len); 83 skb_set_owner_r(skbn, sk); 84 skb_queue_tail(&sk->receive_queue, skbn); 86 sk->data_ready(sk,skbn->len);
|
H A D | x25_link.c | 88 struct sk_buff *skbn; local 114 while ((skbn = skb_dequeue(&neigh->queue)) != NULL) 115 x25_send_frame(skbn, neigh);
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/net/rose/ |
H A D | rose_loopback.c | 44 struct sk_buff *skbn; local 46 skbn = skb_clone(skb, GFP_ATOMIC); 50 if (skbn != NULL) { 51 skb_queue_tail(&loopback_queue, skbn);
|
H A D | rose_dev.c | 104 struct sk_buff *skbn; local 111 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { 117 skb_set_owner_w(skbn, skb->sk); 121 if (!rose_route_frame(skbn, NULL)) { 122 kfree_skb(skbn); 128 stats->tx_bytes += skbn->len;
|
H A D | rose_out.c | 59 struct sk_buff *skb, *skbn; local 87 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { 92 skb_set_owner_w(skbn, sk); 97 rose_send_iframe(sk, skbn);
|
H A D | rose_link.c | 143 struct sk_buff *skbn; local 168 while ((skbn = skb_dequeue(&neigh->queue)) != NULL) 169 if (!rose_send_frame(skbn, neigh)) 170 kfree_skb(skbn);
|
H A D | af_rose.c | 1119 struct sk_buff *skbn; local 1130 if ((skbn = sock_alloc_send_skb(sk, frontlen + ROSE_PACLEN, 0, &err)) == NULL) 1133 skbn->sk = sk; 1134 skbn->free = 1; 1135 skbn->arp = 1; 1137 skb_reserve(skbn, frontlen); 1142 memcpy(skb_put(skbn, lg), skb->data, lg); 1146 skb_push(skbn, ROSE_MIN_LEN); 1147 memcpy(skbn->data, header, ROSE_MIN_LEN); 1150 skbn [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/net/lapb/ |
H A D | lapb_out.c | 73 struct sk_buff *skb, *skbn; local 93 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { 99 skb_set_owner_w(skbn, skb->sk); 104 lapb_send_iframe(lapb, skbn, LAPB_POLLOFF);
|