/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.c | 159 struct bnxt_qplib_nq *nq = nq_work->nq; local 161 if (cq && nq) { 163 if (atomic_read(&cq->arm_state) && nq->cqn_handler) { 164 dev_dbg(&nq->pdev->dev, 165 "%s:Trigger cq = %p event nq = %p\n", 166 __func__, cq, nq); 167 nq->cqn_handler(nq, cq); 235 static void clean_nq(struct bnxt_qplib_nq *nq, struc argument 299 struct bnxt_qplib_nq *nq = from_tasklet(nq, t, nq_tasklet); local 392 bnxt_re_synchronize_nq(struct bnxt_qplib_nq *nq) argument 403 struct bnxt_qplib_nq *nq = dev_instance; local 417 bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill) argument 437 bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq) argument 457 bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx, int msix_vector, bool need_init) argument 498 bnxt_qplib_map_nq_db(struct bnxt_qplib_nq *nq, u32 reg_offt) argument 533 bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq, int nq_idx, int msix_vector, int bar_reg_offset, cqn_handler_t cqn_handler, srqn_handler_t srqn_handler) argument 566 bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq) argument 574 bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq) argument [all...] |
H A D | qplib_fp.h | 420 struct bnxt_qplib_nq *nq; member in struct:bnxt_qplib_cq 483 typedef int (*cqn_handler_t)(struct bnxt_qplib_nq *nq, 485 typedef int (*srqn_handler_t)(struct bnxt_qplib_nq *nq, 508 struct bnxt_qplib_nq *nq; member in struct:bnxt_qplib_nq_work 512 void bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill); 513 void bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq); 514 int bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx, 516 int bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq, 561 void bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq); 562 int bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq); [all...] |
H A D | main.c | 305 struct bnxt_qplib_nq *nq; local 309 nq = &rdev->nq[indx - 1]; 310 bnxt_qplib_nq_stop_irq(nq, false); 321 struct bnxt_qplib_nq *nq; local 347 nq = &rdev->nq[indx - 1]; 348 rc = bnxt_qplib_nq_start_irq(nq, indx - 1, 1187 static int bnxt_re_srqn_handler(struct bnxt_qplib_nq *nq, argument 1206 static int bnxt_re_cqn_handler(struct bnxt_qplib_nq *nq, argument 1315 struct bnxt_qplib_nq *nq; local [all...] |
H A D | bnxt_re.h | 171 struct bnxt_qplib_nq nq[BNXT_RE_MAX_MSIX]; member in struct:bnxt_re_dev
|
H A D | ib_verbs.c | 936 scq_nq = qplib_qp->scq->nq; 937 rcq_nq = qplib_qp->rcq->nq; 1684 struct bnxt_qplib_nq *nq = NULL; local 1687 nq = qplib_srq->cq->nq; 1691 if (nq) 1692 nq->budget--; 1733 struct bnxt_qplib_nq *nq = NULL; local 1776 srq->qplib_srq.eventq_hw_ring_id = rdev->nq[0].ring_id; 1777 nq 2928 struct bnxt_qplib_nq *nq; local 2959 struct bnxt_qplib_nq *nq = NULL; local [all...] |
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 781 ktime_t kt = cmd->nq->dev->completion_nsec; 1208 struct nullb *nullb = cmd->nq->dev->nullb; 1234 struct nullb_device *dev = cmd->nq->dev; 1257 struct badblocks *bb = &cmd->nq->dev->badblocks; 1272 struct nullb_device *dev = cmd->nq->dev; 1283 struct nullb_device *dev = cmd->nq->dev; 1307 switch (cmd->nq->dev->irqmode) { 1323 struct nullb_device *dev = cmd->nq->dev; 1341 struct nullb_device *dev = cmd->nq->dev; 1395 struct nullb_device *dev = cmd->nq 1487 struct nullb_queue *nq = hctx->driver_data; local 1522 struct nullb_queue *nq = hctx->driver_data; local 1554 struct nullb_queue *nq = hctx->driver_data; local 1623 null_init_queue(struct nullb *nullb, struct nullb_queue *nq) argument 1634 struct nullb_queue *nq; local [all...] |
H A D | null_blk.h | 21 struct nullb_queue *nq; member in struct:nullb_cmd
|
H A D | zoned.c | 362 struct nullb_device *dev = cmd->nq->dev; 595 struct nullb_device *dev = cmd->nq->dev; 673 dev = cmd->nq->dev;
|
/linux-master/lib/crypto/ |
H A D | curve25519-hacl64.c | 544 ladder_smallloop_cmult_small_loop_step(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, argument 549 point_swap_conditional(nq, nqpq, bit0); 550 addanddouble_fmonty(nq2, nqpq2, nq, nqpq, q); 556 ladder_smallloop_cmult_small_loop_double_step(u64 *nq, u64 *nqpq, u64 *nq2, argument 560 ladder_smallloop_cmult_small_loop_step(nq, nqpq, nq2, nqpq2, q, byt); 562 ladder_smallloop_cmult_small_loop_step(nq2, nqpq2, nq, nqpq, q, byt1); 566 ladder_smallloop_cmult_small_loop(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, argument 570 ladder_smallloop_cmult_small_loop_double_step(nq, nqpq, nq2, 576 static __always_inline void ladder_bigloop_cmult_big_loop(u8 *n1, u64 *nq, argument 583 ladder_smallloop_cmult_small_loop(nq, nqp 591 u64 *nq = point_buf; local 776 u64 *nq = buf; local [all...] |
/linux-master/fs/xfs/ |
H A D | xfs_trans_dquot.c | 79 struct xfs_dqtrx *oq, *nq; local 98 nq = &nqa[i]; 103 nq->qt_dquot = oq->qt_dquot; 104 nq->qt_bcount_delta = nq->qt_icount_delta = 0; 105 nq->qt_rtbcount_delta = 0; 110 nq->qt_blk_res = oq->qt_blk_res - blk_res_used; 113 nq->qt_rtblk_res = oq->qt_rtblk_res - 117 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used;
|
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 348 struct netdev_queue *nq; local 350 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index); 352 __netif_tx_lock_bh(nq); 353 netif_tx_wake_queue(nq); 354 __netif_tx_unlock_bh(nq); 359 struct netdev_queue *nq; local 362 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index); 364 __netif_tx_lock_bh(nq); 365 netif_tx_stop_queue(nq); 366 __netif_tx_unlock_bh(nq); 814 struct netdev_queue *nq; local 907 struct netdev_queue *nq; local 2307 struct netdev_queue *nq; local [all...] |
/linux-master/drivers/net/vmxnet3/ |
H A D | vmxnet3_xdp.c | 205 struct netdev_queue *nq; local 212 nq = netdev_get_tx_queue(adapter->netdev, tq->qid); 214 __netif_tx_lock(nq, smp_processor_id()); 216 __netif_tx_unlock(nq);
|
/linux-master/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_txrx.c | 427 struct netdev_queue *nq; local 448 nq = netdev_get_tx_queue(netdev, txq->index); 449 __netif_tx_lock(nq, cpu); 450 txq_trans_cond_update(nq); 452 if (netif_tx_queue_stopped(nq) || 456 __netif_tx_unlock(nq); 478 __netif_tx_unlock(nq); 493 struct netdev_queue *nq; local 565 nq = netdev_get_tx_queue(netdev, txq->index); 566 __netif_tx_lock(nq, smp_processor_i [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_uld.c | 111 unsigned int nq = rxq_info->nrxq + rxq_info->nciq; local 125 for (i = 0; i < nq; i++, q++) { 403 int nq = txq_info->ntxq; local 406 for (i = 0; i < nq; i++) { 426 int nq = txq_info->ntxq; local 429 j = nq / adap->params.nports; 430 for (i = 0; i < nq; i++) {
|
/linux-master/arch/x86/crypto/ |
H A D | curve25519-x86_64.c | 977 u64 *nq = p01_tmp1; local 981 u64 *x2 = nq; 982 u64 *z2 = nq + (u32)4U; 1024 fmul2(nq, dc1, ab1, tmp2); 1028 static void point_double(u64 *nq, u64 *tmp1, u64 *tmp2) argument 1030 u64 *x2 = nq; 1031 u64 *z2 = nq + (u32)4U; 1048 fmul2(nq, dc, ab, tmp2);
|
/linux-master/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_singleq_txrx.c | 200 struct netdev_queue *nq; local 288 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx); 289 netdev_tx_sent_queue(nq, first->bytecount); 455 struct netdev_queue *nq; local 559 nq = netdev_get_tx_queue(vport->netdev, tx_q->idx); 563 __netif_txq_completed_wake(nq, total_pkts, total_bytes,
|
H A D | idpf_txrx.c | 1902 struct netdev_queue *nq; local 1912 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx); 1918 __netif_txq_completed_wake(nq, tx_q->cleaned_pkts, tx_q->cleaned_bytes, 1981 struct netdev_queue *nq; local 1990 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx); 1992 return netif_txq_maybe_stop(nq, IDPF_DESC_UNUSED(tx_q), size, size); 2046 struct netdev_queue *nq; local 2048 nq = netdev_get_tx_queue(tx_q->vport->netdev, tx_q->idx); 2061 if (netif_xmit_stopped(nq) || !xmit_more) 2192 struct netdev_queue *nq; local [all...] |
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 1870 struct netdev_queue *nq, bool napi) 1909 netdev_tx_completed_queue(nq, pkts_compl, bytes_compl); 1916 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local 1923 mvneta_txq_bufs_free(pp, txq, tx_done, nq, true); 1927 if (netif_tx_queue_stopped(nq)) { 1929 netif_tx_wake_queue(nq); 2178 struct netdev_queue *nq; local 2189 nq = netdev_get_tx_queue(pp->dev, txq->id); 2191 __netif_tx_lock(nq, cpu); 2206 __netif_tx_unlock(nq); 1868 mvneta_txq_bufs_free(struct mvneta_port *pp, struct mvneta_tx_queue *txq, int num, struct netdev_queue *nq, bool napi) argument 2220 struct netdev_queue *nq; local 2976 struct netdev_queue *nq = netdev_get_tx_queue(dev, txq_id); local 3009 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local 3026 struct netdev_queue *nq; local 3602 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); local [all...] |
H A D | mv643xx_eth.c | 496 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local 498 if (netif_tx_queue_stopped(nq)) { 499 __netif_tx_lock(nq, smp_processor_id()); 501 netif_tx_wake_queue(nq); 502 __netif_tx_unlock(nq); 998 struct netdev_queue *nq; local 1002 nq = netdev_get_tx_queue(dev, queue); 1021 netif_tx_stop_queue(nq); 1035 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local 1039 __netif_tx_lock(nq, smp_processor_i 1060 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); local [all...] |
/linux-master/drivers/net/ethernet/freescale/ |
H A D | fec_main.c | 927 struct netdev_queue *nq; local 932 nq = netdev_get_tx_queue(ndev, queue); 943 netif_tx_stop_queue(nq); 1420 struct netdev_queue *nq; local 1430 nq = netdev_get_tx_queue(ndev, queue_id); 1554 if (netif_tx_queue_stopped(nq)) { 1557 netif_tx_wake_queue(nq); 3915 struct netdev_queue *nq; local 3920 nq = netdev_get_tx_queue(fep->netdev, queue); 3922 __netif_tx_lock(nq, cp 3942 struct netdev_queue *nq; local [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | t4vf_hw.c | 1272 int nq = min(n, 32); local 1277 * "nq" Ingress Queue IDs to the firmware. 1279 cmd.niqid = cpu_to_be16(nq); 1283 * "nq" more done for the start of the next loop. 1285 start += nq; 1286 n -= nq; 1293 while (nq > 0) { 1302 int nqbuf = min(3, nq); 1304 nq -= nqbuf;
|
/linux-master/drivers/net/ |
H A D | tap.c | 192 struct tap_queue *nq; local 203 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]); 204 nq->queue_index = index; 206 rcu_assign_pointer(tap->taps[index], nq);
|
/linux-master/arch/s390/mm/ |
H A D | pgtable.c | 803 unsigned char key, bool nq) 860 page_set_storage_key(paddr, skey, !nq); 884 bool nq, bool mr, bool mc) 903 rc = set_guest_storage_key(current->mm, addr, key, nq); 802 set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, bool nq) argument 882 cond_set_guest_storage_key(struct mm_struct *mm, unsigned long addr, unsigned char key, unsigned char *oldkey, bool nq, bool mr, bool mc) argument
|
/linux-master/drivers/net/ethernet/intel/igc/ |
H A D | igc_main.c | 2467 struct netdev_queue *nq; local 2475 nq = txring_txq(ring); 2477 __netif_tx_lock(nq, cpu); 2479 txq_trans_cond_update(nq); 2481 __netif_tx_unlock(nq); 2548 struct netdev_queue *nq; local 2553 nq = txring_txq(ring); 2555 __netif_tx_lock(nq, cpu); 2557 __netif_tx_unlock(nq); 2880 struct netdev_queue *nq local 6475 struct netdev_queue *nq; local [all...] |
/linux-master/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_nic.c | 837 struct netdev_queue *nq; local 848 nq = netdev_get_tx_queue(ndev, tx_ring->idx); 849 __netif_tx_lock(nq, cpu); 862 __netif_tx_unlock(nq);
|