/linux-master/drivers/isdn/mISDN/ |
H A D | layer2.h | 22 struct sk_buff_head sendq; member in struct:manager
|
H A D | dsp.h | 186 struct sk_buff_head sendq; member in struct:dsp
|
H A D | dsp_core.c | 149 * Send data will be writte to sendq. Sendq will be sent if confirm is received. 887 skb_queue_tail(&dsp->sendq, skb); 934 skb_queue_purge(&dsp->sendq); 978 skb_queue_purge(&dsp->sendq); 1014 while ((skb = skb_dequeue(&dsp->sendq))) { 1065 skb_queue_head_init(&ndsp->sendq);
|
H A D | tei.c | 335 struct sk_buff *skb = skb_dequeue(&mgr->sendq); 359 skb = skb_dequeue(&mgr->sendq); 376 skb_queue_tail(&mgr->sendq, skb); 399 skb_queue_tail(&mgr->sendq, skb); 1253 skb_queue_purge(&mgr->sendq); 1353 skb_queue_head_init(&mgr->sendq);
|
H A D | dsp_cmx.c | 1573 skb_queue_tail(&dsp->sendq, nskb); 1589 skb_queue_tail(&dsp->sendq, txskb); 1606 skb_queue_tail(&dsp->sendq, nskb); 1928 skb_queue_tail(&dsp->sendq, nskb); 1944 skb_queue_tail(&member->dsp->sendq, nskb);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 2747 spin_lock(&q->sendq.lock); 2751 __skb_queue_tail(&q->sendq, skb); 2752 spin_unlock(&q->sendq.lock); 2764 spin_unlock(&q->sendq.lock); 2782 spin_lock(&q->sendq.lock); 2786 while ((skb = __skb_dequeue(&q->sendq)) != NULL) { 2796 spin_unlock(&q->sendq.lock); 2806 spin_lock(&q->sendq.lock); 2814 spin_lock(&q->sendq.lock); 2820 spin_unlock(&q->sendq [all...] |
H A D | cxgb4.h | 882 struct sk_buff_head sendq; /* list of backpressured packets */ member in struct:sge_uld_txq 884 bool service_ofldq_running; /* service_ofldq() is processing sendq */ 892 struct sk_buff_head sendq; /* list of backpressured packets */ member in struct:sge_ctrl_txq
|
H A D | cxgb4_uld.c | 415 __skb_queue_purge(&txq->sendq);
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw.h | 436 struct siw_sqe *sendq; /* send queue element array */ member in struct:siw_qp 628 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; 635 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size];
|
H A D | siw_verbs.c | 380 qp->sendq = vmalloc_user(num_sqe * sizeof(struct siw_sqe)); 382 qp->sendq = vcalloc(num_sqe, sizeof(struct siw_sqe)); 384 if (qp->sendq == NULL) { 440 if (qp->sendq) { 443 siw_mmap_entry_insert(uctx, qp->sendq, 491 vfree(qp->sendq); 818 struct siw_sqe *sqe = &qp->sendq[idx];
|
H A D | siw_qp.c | 1237 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; 1342 vfree(qp->sendq);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | adapter.h | 180 struct sk_buff_head sendq; /* List of backpressured offload packets */ member in struct:sge_txq
|
H A D | sge.c | 704 __skb_queue_purge(&q->txq[i].sendq); 1422 if (unlikely(!skb_queue_empty(&q->sendq))) { 1423 addq_exit:__skb_queue_tail(&q->sendq, skb); 1531 (skb = __skb_dequeue(&q->sendq)) != NULL) { 1542 if (!skb_queue_empty(&q->sendq)) { 1753 while ((skb = skb_peek(&q->sendq)) != NULL) { 1781 __skb_unlink(skb, &q->sendq); 3070 skb_queue_head_init(&q->txq[i].sendq);
|
/linux-master/drivers/rpmsg/ |
H A D | virtio_rpmsg_bus.c | 49 * @sendq: wait queue of sending contexts waiting for a tx buffers 67 wait_queue_head_t sendq; member in struct:virtproc_info 600 err = wait_event_interruptible_timeout(vrp->sendq, 823 wake_up_interruptible(&vrp->sendq); 891 init_waitqueue_head(&vrp->sendq);
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci.c | 712 u8 sendq = mlxsw_pci_cqe_sr_get(q->u.cq.v, cqe); local 715 if (unlikely(sendq)) { 770 u8 sendq = mlxsw_pci_cqe_sr_get(q->u.cq.v, cqe); local 774 if (unlikely(!sendq)) {
|
/linux-master/drivers/crypto/chelsio/ |
H A D | chcr_algo.c | 585 spin_lock(&txq->sendq.lock); 588 spin_unlock(&txq->sendq.lock);
|