Searched refs:q_idx (Results 1 - 25 of 63) sorted by path

123

/linux-master/drivers/accel/habanalabs/common/
H A Dcommand_submission.c155 "SOB release shouldn't be called here, q_idx: %d, sob_id: %d\n",
156 hw_sob->q_idx, hw_sob->sob_id);
1793 * @q_idx: stream queue index
1800 int hl_cs_signal_sob_wraparound_handler(struct hl_device *hdev, u32 q_idx, argument
1808 prop = &hdev->kernel_queues[q_idx].sync_stream_prop;
1830 dev_err(hdev->dev, "error: Cannot switch SOBs q_idx: %d\n",
1831 q_idx);
1871 dev_dbg(hdev->dev, "switched to SOB %d, q_idx: %d\n",
1872 prop->curr_sob_offset, q_idx);
1940 enum hl_queue_type q_type, u32 q_idx, u3
1938 cs_ioctl_signal_wait_create_jobs(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_cs *cs, enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) argument
2003 cs_ioctl_reserve_signals(struct hl_fpriv *hpriv, u32 q_idx, u32 count, u32 *handle_id, u32 *sob_addr, u32 *signals_count) argument
2132 u32 q_idx, sob_addr; local
2212 u32 q_idx, collective_engine_id = 0; local
[all...]
H A Dhabanalabs.h261 * @q_idx: queue id is used to extract fence register address
270 u32 q_idx; member in struct:hl_gen_wait_properties
420 * @q_idx: the H/W queue that uses this SOB.
428 u32 q_idx; member in struct:hl_hw_sob
3545 * @q_idx: stream queue index
3556 u32 q_idx; member in struct:hl_cs_encaps_sig_handle
4022 int hl_cs_signal_sob_wraparound_handler(struct hl_device *hdev, u32 q_idx,
H A Dhw_queue.c420 u32 q_idx; local
423 q_idx = job->hw_queue_id;
424 prop = &hdev->kernel_queues[q_idx].sync_stream_prop;
431 "generate signal CB, sob_id: %d, sob val: %u, q_idx: %d, seq: %llu\n",
432 cs_cmpl->hw_sob->sob_id, cs_cmpl->sob_val, q_idx,
441 rc = hl_cs_signal_sob_wraparound_handler(hdev, q_idx, &hw_sob, 1,
480 u32 q_idx; local
482 q_idx = job->hw_queue_id;
483 prop = &hdev->kernel_queues[q_idx].sync_stream_prop;
497 cs->encaps_sig_hdl->q_idx,
913 sync_stream_queue_init(struct hl_device *hdev, u32 q_idx) argument
972 sync_stream_queue_reset(struct hl_device *hdev, u32 q_idx) argument
[all...]
/linux-master/drivers/accel/habanalabs/gaudi/
H A Dgaudi.c1240 wait_prop.q_idx = queue_id;
1290 wait_prop.q_idx = queue_id;
1301 "generate signal CB, sob_id: %d, sob val: 1, q_idx: %d\n",
2690 int i, j, dma_id, cpu_skip, nic_skip, cq_id = 0, q_idx, msi_vec = 0; local
2711 q_idx = 4 * dma_id + j + cpu_skip;
2712 q = &hdev->kernel_queues[q_idx];
8739 if (gaudi_get_fence_addr(hdev, prop->q_idx, &fence_addr)) {
8741 prop->q_idx);
8757 dev_dbg(hdev->dev, "reset SOB, q_idx: %d, sob_id: %d\n", hw_sob->q_idx,
[all...]
/linux-master/drivers/accel/habanalabs/gaudi2/
H A Dgaudi2.c11114 stream_index = prop->q_idx % 4;
11115 fence_addr = CFG_BASE + gaudi2_qm_blocks_bases[prop->q_idx] +
11161 dev_dbg(hdev->dev, "reset SOB, q_idx: %d, sob_id: %d\n", hw_sob->q_idx, hw_sob->sob_id);
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c2883 int q_idx = QUEUE_NOT_FOUND; local
2886 q_idx = q_array_get_index(
2891 if (!queue_ids || q_idx != QUEUE_NOT_FOUND) {
2896 queue_ids[q_idx] &=
2899 queue_ids[q_idx] |=
2928 int q_idx = q_array_get_index( local
2934 if (q_idx != QUEUE_NOT_FOUND)
2935 queue_ids[q_idx] |=
2987 int q_idx = q_array_get_index(q->properties.queue_id, local
2991 if (q_idx !
3031 int q_idx = q_array_get_index(q->properties.queue_id, local
[all...]
/linux-master/drivers/infiniband/hw/hfi1/
H A Dhfi.h1301 int (*process_vnic_dma_send)(struct hfi1_devdata *dd, u8 q_idx,
H A Dipoib.h102 * @q_idx: ipoib Tx queue index
114 u8 q_idx; member in struct:hfi1_ipoib_txq
H A Dipoib_tx.c57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx);
64 netif_wake_subqueue(txq->priv->netdev, txq->q_idx);
125 le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx,
723 txq->q_idx = i;
789 txq->q_idx,
848 __netif_subqueue_stopped(dev, txq->q_idx),
H A Dtrace_tx.h918 __entry->idx = txq->q_idx;
926 __netif_subqueue_stopped(txq->priv->netdev, txq->q_idx);
1023 __entry->txq ? __entry->txq->q_idx : 0,
H A Dvnic.h40 * @q_idx - vnic Tx queue index
49 u8 q_idx; member in struct:hfi1_vnic_sdma
113 u8 q_idx);
122 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx,
H A Dvnic_main.c124 u8 q_idx, struct sk_buff *skb, int err)
127 struct opa_vnic_stats *stats = &vinfo->stats[q_idx];
153 u8 q_idx, struct sk_buff *skb, int err)
156 struct opa_vnic_stats *stats = &vinfo->stats[q_idx];
205 u8 q_idx)
207 netif_stop_subqueue(vinfo->netdev, q_idx);
208 if (!hfi1_vnic_sdma_write_avail(vinfo, q_idx))
211 netif_start_subqueue(vinfo->netdev, q_idx);
218 u8 pad_len, q_idx = skb->queue_mapping; local
225 v_dbg("xmit: queue %d skb len %d\n", q_idx, sk
123 hfi1_vnic_update_tx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument
152 hfi1_vnic_update_rx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument
204 hfi1_vnic_maybe_stop_tx(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument
339 u8 q_idx; local
[all...]
H A Dvnic_sdma.c126 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, argument
130 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
224 if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx))
225 netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx);
229 u8 q_idx)
231 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx];
249 vnic_sdma->q_idx = i;
228 hfi1_vnic_sdma_write_avail(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument
/linux-master/drivers/net/ethernet/broadcom/bnx2x/
H A Dbnx2x_sriov.c129 u16 q_idx, u16 sb_idx)
134 q_idx,
145 u16 q_idx, u16 sb_idx)
152 q_idx,
126 bnx2x_vfop_qctor_dump_tx(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_queue_init_params *init_params, struct bnx2x_queue_setup_params *setup_params, u16 q_idx, u16 sb_idx) argument
142 bnx2x_vfop_qctor_dump_rx(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_queue_init_params *init_params, struct bnx2x_queue_setup_params *setup_params, u16 q_idx, u16 sb_idx) argument
H A Dbnx2x_sriov.h438 u16 q_idx, u16 sb_idx);
443 u16 q_idx, u16 sb_idx);
/linux-master/drivers/net/ethernet/cavium/liquidio/
H A Dlio_main.c2313 int q_idx = 0, iq_no = 0; local
2321 q_idx = skb_iq(oct, skb);
2322 tag = q_idx;
2323 iq_no = lio->linfo.txpciq[q_idx].s.q_no;
2412 spin_lock(&lio->glist_lock[q_idx]);
2414 lio_list_delete_head(&lio->glist[q_idx]);
2415 spin_unlock(&lio->glist_lock[q_idx]);
2518 netif_stop_subqueue(netdev, q_idx);
H A Dlio_vf_main.c1414 int q_idx = 0, iq_no = 0; local
1426 q_idx = skb_iq(lio->oct_dev, skb);
1427 tag = q_idx;
1428 iq_no = lio->linfo.txpciq[q_idx].s.q_no;
1506 spin_lock(&lio->glist_lock[q_idx]);
1508 lio_list_delete_head(&lio->glist[q_idx]);
1509 spin_unlock(&lio->glist_lock[q_idx]);
1603 netif_stop_subqueue(netdev, q_idx);
/linux-master/drivers/net/ethernet/cavium/thunder/
H A Dnicvf_queues.c1723 static u64 nicvf_int_type_to_mask(int int_type, int q_idx) argument
1729 reg_val = ((1ULL << q_idx) << NICVF_INTR_CQ_SHIFT);
1732 reg_val = ((1ULL << q_idx) << NICVF_INTR_SQ_SHIFT);
1735 reg_val = ((1ULL << q_idx) << NICVF_INTR_RBDR_SHIFT);
1757 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx) argument
1759 u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
1771 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx) argument
1773 u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
1785 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx) argument
1787 u64 mask = nicvf_int_type_to_mask(int_type, q_idx);
1799 nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx) argument
[all...]
H A Dnicvf_queues.h354 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx);
355 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx);
356 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx);
357 int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx);
/linux-master/drivers/net/ethernet/chelsio/cxgb3/
H A Dt3_hw.c2473 int i, j, cpu_idx = 0, q_idx = 0; local
2490 (i << 16) | rspq[q_idx++]);
2491 if (rspq[q_idx] == 0xffff)
2492 q_idx = 0;
/linux-master/drivers/net/ethernet/emulex/benet/
H A Dbe_main.c1371 u16 q_idx = skb_get_queue_mapping(skb); local
1372 struct be_tx_obj *txo = &adapter->tx_obj[q_idx];
1402 netif_stop_subqueue(netdev, q_idx);
1406 if (flush || __netif_subqueue_stopped(netdev, q_idx))
/linux-master/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_tx.c576 int i, q_idx; local
585 q_idx = smp_processor_id();
586 if (unlikely(q_idx >= fp->num_xdpqs))
589 for (q = xdpqs[q_idx], i = 0; i < n; i++)
/linux-master/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_main.c1874 int q_idx = pc; local
1877 interface->tx_ring[offset + i]->reg_idx = q_idx;
1879 interface->rx_ring[offset + i]->reg_idx = q_idx;
1881 q_idx += pc_stride;
H A Dfm10k_pci.c1175 int q_idx; local
1177 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) {
1178 q_vector = interface->q_vector[q_idx];
1872 int q_idx; local
1874 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) {
1875 q_vector = interface->q_vector[q_idx];
[all...]
H A Dfm10k_pf.c503 u16 vsi, queue, pc, q_idx; local
520 q_idx = dglort->queue_b;
524 for (queue = 0; queue < queue_count; queue++, q_idx++) {
525 if (q_idx >= FM10K_MAX_QUEUES)
528 fm10k_write_reg(hw, FM10K_TX_SGLORT(q_idx), glort);
529 fm10k_write_reg(hw, FM10K_RX_SGLORT(q_idx), glort);
539 q_idx = pc + dglort->queue_b;
541 if (q_idx >= FM10K_MAX_QUEUES)
544 txqctl = fm10k_read_reg(hw, FM10K_TXQCTL(q_idx));
547 fm10k_write_reg(hw, FM10K_TXQCTL(q_idx), txqct
[all...]

Completed in 520 milliseconds

123