Searched refs:nq (Results 1 - 25 of 37) sorted by relevance

12

/linux-master/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c159 struct bnxt_qplib_nq *nq = nq_work->nq; local
161 if (cq && nq) {
163 if (atomic_read(&cq->arm_state) && nq->cqn_handler) {
164 dev_dbg(&nq->pdev->dev,
165 "%s:Trigger cq = %p event nq = %p\n",
166 __func__, cq, nq);
167 nq->cqn_handler(nq, cq);
235 static void clean_nq(struct bnxt_qplib_nq *nq, struc argument
299 struct bnxt_qplib_nq *nq = from_tasklet(nq, t, nq_tasklet); local
392 bnxt_re_synchronize_nq(struct bnxt_qplib_nq *nq) argument
403 struct bnxt_qplib_nq *nq = dev_instance; local
417 bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill) argument
437 bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq) argument
457 bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx, int msix_vector, bool need_init) argument
498 bnxt_qplib_map_nq_db(struct bnxt_qplib_nq *nq, u32 reg_offt) argument
533 bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq, int nq_idx, int msix_vector, int bar_reg_offset, cqn_handler_t cqn_handler, srqn_handler_t srqn_handler) argument
566 bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq) argument
574 bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq) argument
[all...]
H A Dqplib_fp.h420 struct bnxt_qplib_nq *nq; member in struct:bnxt_qplib_cq
483 typedef int (*cqn_handler_t)(struct bnxt_qplib_nq *nq,
485 typedef int (*srqn_handler_t)(struct bnxt_qplib_nq *nq,
508 struct bnxt_qplib_nq *nq; member in struct:bnxt_qplib_nq_work
512 void bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill);
513 void bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq);
514 int bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx,
516 int bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq,
561 void bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq);
562 int bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq);
[all...]
H A Dmain.c305 struct bnxt_qplib_nq *nq; local
309 nq = &rdev->nq[indx - 1];
310 bnxt_qplib_nq_stop_irq(nq, false);
321 struct bnxt_qplib_nq *nq; local
347 nq = &rdev->nq[indx - 1];
348 rc = bnxt_qplib_nq_start_irq(nq, indx - 1,
1187 static int bnxt_re_srqn_handler(struct bnxt_qplib_nq *nq, argument
1206 static int bnxt_re_cqn_handler(struct bnxt_qplib_nq *nq, argument
1315 struct bnxt_qplib_nq *nq; local
[all...]
H A Dbnxt_re.h171 struct bnxt_qplib_nq nq[BNXT_RE_MAX_MSIX]; member in struct:bnxt_re_dev
H A Dib_verbs.c936 scq_nq = qplib_qp->scq->nq;
937 rcq_nq = qplib_qp->rcq->nq;
1684 struct bnxt_qplib_nq *nq = NULL; local
1687 nq = qplib_srq->cq->nq;
1691 if (nq)
1692 nq->budget--;
1733 struct bnxt_qplib_nq *nq = NULL; local
1776 srq->qplib_srq.eventq_hw_ring_id = rdev->nq[0].ring_id;
1777 nq
2928 struct bnxt_qplib_nq *nq; local
2959 struct bnxt_qplib_nq *nq = NULL; local
[all...]
/linux-master/drivers/block/null_blk/
H A Dmain.c781 ktime_t kt = cmd->nq->dev->completion_nsec;
1208 struct nullb *nullb = cmd->nq->dev->nullb;
1234 struct nullb_device *dev = cmd->nq->dev;
1257 struct badblocks *bb = &cmd->nq->dev->badblocks;
1272 struct nullb_device *dev = cmd->nq->dev;
1283 struct nullb_device *dev = cmd->nq->dev;
1307 switch (cmd->nq->dev->irqmode) {
1323 struct nullb_device *dev = cmd->nq->dev;
1341 struct nullb_device *dev = cmd->nq->dev;
1395 struct nullb_device *dev = cmd->nq
1487 struct nullb_queue *nq = hctx->driver_data; local
1522 struct nullb_queue *nq = hctx->driver_data; local
1554 struct nullb_queue *nq = hctx->driver_data; local
1623 null_init_queue(struct nullb *nullb, struct nullb_queue *nq) argument
1634 struct nullb_queue *nq; local
[all...]
H A Dnull_blk.h21 struct nullb_queue *nq; member in struct:nullb_cmd
H A Dzoned.c362 struct nullb_device *dev = cmd->nq->dev;
595 struct nullb_device *dev = cmd->nq->dev;
673 dev = cmd->nq->dev;
/linux-master/lib/crypto/
H A Dcurve25519-hacl64.c544 ladder_smallloop_cmult_small_loop_step(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, argument
549 point_swap_conditional(nq, nqpq, bit0);
550 addanddouble_fmonty(nq2, nqpq2, nq, nqpq, q);
556 ladder_smallloop_cmult_small_loop_double_step(u64 *nq, u64 *nqpq, u64 *nq2, argument
560 ladder_smallloop_cmult_small_loop_step(nq, nqpq, nq2, nqpq2, q, byt);
562 ladder_smallloop_cmult_small_loop_step(nq2, nqpq2, nq, nqpq, q, byt1);
566 ladder_smallloop_cmult_small_loop(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, argument
570 ladder_smallloop_cmult_small_loop_double_step(nq, nqpq, nq2,
576 static __always_inline void ladder_bigloop_cmult_big_loop(u8 *n1, u64 *nq, argument
583 ladder_smallloop_cmult_small_loop(nq, nqp
591 u64 *nq = point_buf; local
776 u64 *nq = buf; local
[all...]
/linux-master/fs/xfs/
H A Dxfs_trans_dquot.c79 struct xfs_dqtrx *oq, *nq; local
98 nq = &nqa[i];
103 nq->qt_dquot = oq->qt_dquot;
104 nq->qt_bcount_delta = nq->qt_icount_delta = 0;
105 nq->qt_rtbcount_delta = 0;
110 nq->qt_blk_res = oq->qt_blk_res - blk_res_used;
113 nq->qt_rtblk_res = oq->qt_rtblk_res -
117 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used;
/linux-master/drivers/net/ethernet/engleder/
H A Dtsnep_main.c348 struct netdev_queue *nq; local
350 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
352 __netif_tx_lock_bh(nq);
353 netif_tx_wake_queue(nq);
354 __netif_tx_unlock_bh(nq);
359 struct netdev_queue *nq; local
362 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
364 __netif_tx_lock_bh(nq);
365 netif_tx_stop_queue(nq);
366 __netif_tx_unlock_bh(nq);
814 struct netdev_queue *nq; local
907 struct netdev_queue *nq; local
2307 struct netdev_queue *nq; local
[all...]
/linux-master/drivers/net/vmxnet3/
H A Dvmxnet3_xdp.c205 struct netdev_queue *nq; local
212 nq = netdev_get_tx_queue(adapter->netdev, tq->qid);
214 __netif_tx_lock(nq, smp_processor_id());
216 __netif_tx_unlock(nq);
/linux-master/drivers/net/ethernet/pensando/ionic/
H A Dionic_txrx.c427 struct netdev_queue *nq; local
448 nq = netdev_get_tx_queue(netdev, txq->index);
449 __netif_tx_lock(nq, cpu);
450 txq_trans_cond_update(nq);
452 if (netif_tx_queue_stopped(nq) ||
456 __netif_tx_unlock(nq);
478 __netif_tx_unlock(nq);
493 struct netdev_queue *nq; local
565 nq = netdev_get_tx_queue(netdev, txq->index);
566 __netif_tx_lock(nq, smp_processor_i
[all...]
/linux-master/drivers/net/ethernet/chelsio/cxgb4/
H A Dcxgb4_uld.c111 unsigned int nq = rxq_info->nrxq + rxq_info->nciq; local
125 for (i = 0; i < nq; i++, q++) {
403 int nq = txq_info->ntxq; local
406 for (i = 0; i < nq; i++) {
426 int nq = txq_info->ntxq; local
429 j = nq / adap->params.nports;
430 for (i = 0; i < nq; i++) {
/linux-master/arch/x86/crypto/
H A Dcurve25519-x86_64.c977 u64 *nq = p01_tmp1; local
981 u64 *x2 = nq;
982 u64 *z2 = nq + (u32)4U;
1024 fmul2(nq, dc1, ab1, tmp2);
1028 static void point_double(u64 *nq, u64 *tmp1, u64 *tmp2) argument
1030 u64 *x2 = nq;
1031 u64 *z2 = nq + (u32)4U;
1048 fmul2(nq, dc, ab, tmp2);
/linux-master/drivers/net/ethernet/intel/idpf/
H A Didpf_singleq_txrx.c200 struct netdev_queue *nq; local
288 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx);
289 netdev_tx_sent_queue(nq, first->bytecount);
455 struct netdev_queue *nq; local
559 nq = netdev_get_tx_queue(vport->netdev, tx_q->idx);
563 __netif_txq_completed_wake(nq, total_pkts, total_bytes,
H A Didpf_txrx.c1902 struct netdev_queue *nq; local
1912 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx);
1918 __netif_txq_completed_wake(nq, tx_q->cleaned_pkts, tx_q->cleaned_bytes,
1981 struct netdev_queue *nq; local
1990 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx);
1992 return netif_txq_maybe_stop(nq, IDPF_DESC_UNUSED(tx_q), size, size);
2046 struct netdev_queue *nq; local
2048 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx);
2061 if (netif_xmit_stopped(nq) || !xmit_more)
2192 struct netdev_queue *nq; local
[all...]
/linux-master/drivers/net/ethernet/marvell/
H A Dmvneta.c1870 struct netdev_queue *nq, bool napi)
1909 netdev_tx_completed_queue(nq, pkts_compl, bytes_compl);
1916 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local
1923 mvneta_txq_bufs_free(pp, txq, tx_done, nq, true);
1927 if (netif_tx_queue_stopped(nq)) {
1929 netif_tx_wake_queue(nq);
2178 struct netdev_queue *nq; local
2189 nq = netdev_get_tx_queue(pp->dev, txq->id);
2191 __netif_tx_lock(nq, cpu);
2206 __netif_tx_unlock(nq);
1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) argument
2220 struct netdev_queue *nq; local
2976 struct netdev_queue *nq = netdev_get_tx_queue(dev, txq_id); local
3009 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local
3026 struct netdev_queue *nq; local
3602 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local
[all...]
H A Dmv643xx_eth.c496 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local
498 if (netif_tx_queue_stopped(nq)) {
499 __netif_tx_lock(nq, smp_processor_id());
501 netif_tx_wake_queue(nq);
502 __netif_tx_unlock(nq);
998 struct netdev_queue *nq; local
1002 nq = netdev_get_tx_queue(dev, queue);
1021 netif_tx_stop_queue(nq);
1035 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local
1039 __netif_tx_lock(nq, smp_processor_i
1060 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local
[all...]
/linux-master/drivers/net/ethernet/freescale/
H A Dfec_main.c927 struct netdev_queue *nq; local
932 nq = netdev_get_tx_queue(ndev, queue);
943 netif_tx_stop_queue(nq);
1420 struct netdev_queue *nq; local
1430 nq = netdev_get_tx_queue(ndev, queue_id);
1554 if (netif_tx_queue_stopped(nq)) {
1557 netif_tx_wake_queue(nq);
3915 struct netdev_queue *nq; local
3920 nq = netdev_get_tx_queue(fep->netdev, queue);
3922 __netif_tx_lock(nq, cp
3942 struct netdev_queue *nq; local
[all...]
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/
H A Dt4vf_hw.c1272 int nq = min(n, 32); local
1277 * "nq" Ingress Queue IDs to the firmware.
1279 cmd.niqid = cpu_to_be16(nq);
1283 * "nq" more done for the start of the next loop.
1285 start += nq;
1286 n -= nq;
1293 while (nq > 0) {
1302 int nqbuf = min(3, nq);
1304 nq -= nqbuf;
/linux-master/drivers/net/
H A Dtap.c192 struct tap_queue *nq; local
203 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]);
204 nq->queue_index = index;
206 rcu_assign_pointer(tap->taps[index], nq);
/linux-master/arch/s390/mm/
H A Dpgtable.c803 unsigned char key, bool nq)
860 page_set_storage_key(paddr, skey, !nq);
884 bool nq, bool mr, bool mc)
903 rc = set_guest_storage_key(current->mm, addr, key, nq);
802 set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, bool nq) argument
882 cond_set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, unsigned char *oldkey, bool nq, bool mr, bool mc) argument
/linux-master/drivers/net/ethernet/intel/igc/
H A Digc_main.c2467 struct netdev_queue *nq; local
2475 nq = txring_txq(ring);
2477 __netif_tx_lock(nq, cpu);
2479 txq_trans_cond_update(nq);
2481 __netif_tx_unlock(nq);
2548 struct netdev_queue *nq; local
2553 nq = txring_txq(ring);
2555 __netif_tx_lock(nq, cpu);
2557 __netif_tx_unlock(nq);
2880 struct netdev_queue *nq local
6475 struct netdev_queue *nq; local
[all...]
/linux-master/drivers/net/ethernet/aquantia/atlantic/
H A Daq_nic.c837 struct netdev_queue *nq; local
848 nq = netdev_get_tx_queue(ndev, tx_ring->idx);
849 __netif_tx_lock(nq, cpu);
862 __netif_tx_unlock(nq);

Completed in 484 milliseconds

12