/linux-master/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 155 "SOB release shouldn't be called here, q_idx: %d, sob_id: %d\n", 156 hw_sob->q_idx, hw_sob->sob_id); 1793 * @q_idx: stream queue index 1800 int hl_cs_signal_sob_wraparound_handler(struct hl_device *hdev, u32 q_idx, argument 1808 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 1830 dev_err(hdev->dev, "error: Cannot switch SOBs q_idx: %d\n", 1831 q_idx); 1871 dev_dbg(hdev->dev, "switched to SOB %d, q_idx: %d\n", 1872 prop->curr_sob_offset, q_idx); 1940 enum hl_queue_type q_type, u32 q_idx, u3 1938 cs_ioctl_signal_wait_create_jobs(struct hl_device *hdev, struct hl_ctx *ctx, struct hl_cs *cs, enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) argument 2003 cs_ioctl_reserve_signals(struct hl_fpriv *hpriv, u32 q_idx, u32 count, u32 *handle_id, u32 *sob_addr, u32 *signals_count) argument 2132 u32 q_idx, sob_addr; local 2212 u32 q_idx, collective_engine_id = 0; local [all...] |
H A D | habanalabs.h | 261 * @q_idx: queue id is used to extract fence register address 270 u32 q_idx; member in struct:hl_gen_wait_properties 420 * @q_idx: the H/W queue that uses this SOB. 428 u32 q_idx; member in struct:hl_hw_sob 3545 * @q_idx: stream queue index 3556 u32 q_idx; member in struct:hl_cs_encaps_sig_handle 4022 int hl_cs_signal_sob_wraparound_handler(struct hl_device *hdev, u32 q_idx,
|
H A D | hw_queue.c | 420 u32 q_idx; local 423 q_idx = job->hw_queue_id; 424 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 431 "generate signal CB, sob_id: %d, sob val: %u, q_idx: %d, seq: %llu\n", 432 cs_cmpl->hw_sob->sob_id, cs_cmpl->sob_val, q_idx, 441 rc = hl_cs_signal_sob_wraparound_handler(hdev, q_idx, &hw_sob, 1, 480 u32 q_idx; local 482 q_idx = job->hw_queue_id; 483 prop = &hdev->kernel_queues[q_idx].sync_stream_prop; 497 cs->encaps_sig_hdl->q_idx, 913 sync_stream_queue_init(struct hl_device *hdev, u32 q_idx) argument 972 sync_stream_queue_reset(struct hl_device *hdev, u32 q_idx) argument [all...] |
/linux-master/drivers/accel/habanalabs/gaudi/ |
H A D | gaudi.c | 1240 wait_prop.q_idx = queue_id; 1290 wait_prop.q_idx = queue_id; 1301 "generate signal CB, sob_id: %d, sob val: 1, q_idx: %d\n", 2690 int i, j, dma_id, cpu_skip, nic_skip, cq_id = 0, q_idx, msi_vec = 0; local 2711 q_idx = 4 * dma_id + j + cpu_skip; 2712 q = &hdev->kernel_queues[q_idx]; 8739 if (gaudi_get_fence_addr(hdev, prop->q_idx, &fence_addr)) { 8741 prop->q_idx); 8757 dev_dbg(hdev->dev, "reset SOB, q_idx: %d, sob_id: %d\n", hw_sob->q_idx, [all...] |
/linux-master/drivers/accel/habanalabs/gaudi2/ |
H A D | gaudi2.c | 11114 stream_index = prop->q_idx % 4; 11115 fence_addr = CFG_BASE + gaudi2_qm_blocks_bases[prop->q_idx] + 11161 dev_dbg(hdev->dev, "reset SOB, q_idx: %d, sob_id: %d\n", hw_sob->q_idx, hw_sob->sob_id);
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 2883 int q_idx = QUEUE_NOT_FOUND; local 2886 q_idx = q_array_get_index( 2891 if (!queue_ids || q_idx != QUEUE_NOT_FOUND) { 2896 queue_ids[q_idx] &= 2899 queue_ids[q_idx] |= 2928 int q_idx = q_array_get_index( local 2934 if (q_idx != QUEUE_NOT_FOUND) 2935 queue_ids[q_idx] |= 2987 int q_idx = q_array_get_index(q->properties.queue_id, local 2991 if (q_idx ! 3031 int q_idx = q_array_get_index(q->properties.queue_id, local [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | hfi.h | 1301 int (*process_vnic_dma_send)(struct hfi1_devdata *dd, u8 q_idx,
|
H A D | ipoib.h | 102 * @q_idx: ipoib Tx queue index 114 u8 q_idx; member in struct:hfi1_ipoib_txq
|
H A D | ipoib_tx.c | 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); 64 netif_wake_subqueue(txq->priv->netdev, txq->q_idx); 125 le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx, 723 txq->q_idx = i; 789 txq->q_idx, 848 __netif_subqueue_stopped(dev, txq->q_idx),
|
H A D | trace_tx.h | 918 __entry->idx = txq->q_idx; 926 __netif_subqueue_stopped(txq->priv->netdev, txq->q_idx); 1023 __entry->txq ? __entry->txq->q_idx : 0,
|
H A D | vnic.h | 40 * @q_idx - vnic Tx queue index 49 u8 q_idx; member in struct:hfi1_vnic_sdma 113 u8 q_idx); 122 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx,
|
H A D | vnic_main.c | 124 u8 q_idx, struct sk_buff *skb, int err) 127 struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; 153 u8 q_idx, struct sk_buff *skb, int err) 156 struct opa_vnic_stats *stats = &vinfo->stats[q_idx]; 205 u8 q_idx) 207 netif_stop_subqueue(vinfo->netdev, q_idx); 208 if (!hfi1_vnic_sdma_write_avail(vinfo, q_idx)) 211 netif_start_subqueue(vinfo->netdev, q_idx); 218 u8 pad_len, q_idx = skb->queue_mapping; local 225 v_dbg("xmit: queue %d skb len %d\n", q_idx, sk 123 hfi1_vnic_update_tx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument 152 hfi1_vnic_update_rx_counters(struct hfi1_vnic_vport_info *vinfo, u8 q_idx, struct sk_buff *skb, int err) argument 204 hfi1_vnic_maybe_stop_tx(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument 339 u8 q_idx; local [all...] |
H A D | vnic_sdma.c | 126 int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, argument 130 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; 224 if (__netif_subqueue_stopped(vinfo->netdev, vnic_sdma->q_idx)) 225 netif_wake_subqueue(vinfo->netdev, vnic_sdma->q_idx); 229 u8 q_idx) 231 struct hfi1_vnic_sdma *vnic_sdma = &vinfo->sdma[q_idx]; 249 vnic_sdma->q_idx = i; 228 hfi1_vnic_sdma_write_avail(struct hfi1_vnic_vport_info *vinfo, u8 q_idx) argument
|
/linux-master/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_sriov.c | 129 u16 q_idx, u16 sb_idx) 134 q_idx, 145 u16 q_idx, u16 sb_idx) 152 q_idx, 126 bnx2x_vfop_qctor_dump_tx(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_queue_init_params *init_params, struct bnx2x_queue_setup_params *setup_params, u16 q_idx, u16 sb_idx) argument 142 bnx2x_vfop_qctor_dump_rx(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_queue_init_params *init_params, struct bnx2x_queue_setup_params *setup_params, u16 q_idx, u16 sb_idx) argument
|
H A D | bnx2x_sriov.h | 438 u16 q_idx, u16 sb_idx); 443 u16 q_idx, u16 sb_idx);
|
/linux-master/drivers/net/ethernet/cavium/liquidio/ |
H A D | lio_main.c | 2313 int q_idx = 0, iq_no = 0; local 2321 q_idx = skb_iq(oct, skb); 2322 tag = q_idx; 2323 iq_no = lio->linfo.txpciq[q_idx].s.q_no; 2412 spin_lock(&lio->glist_lock[q_idx]); 2414 lio_list_delete_head(&lio->glist[q_idx]); 2415 spin_unlock(&lio->glist_lock[q_idx]); 2518 netif_stop_subqueue(netdev, q_idx);
|
H A D | lio_vf_main.c | 1414 int q_idx = 0, iq_no = 0; local 1426 q_idx = skb_iq(lio->oct_dev, skb); 1427 tag = q_idx; 1428 iq_no = lio->linfo.txpciq[q_idx].s.q_no; 1506 spin_lock(&lio->glist_lock[q_idx]); 1508 lio_list_delete_head(&lio->glist[q_idx]); 1509 spin_unlock(&lio->glist_lock[q_idx]); 1603 netif_stop_subqueue(netdev, q_idx);
|
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_queues.c | 1723 static u64 nicvf_int_type_to_mask(int int_type, int q_idx) argument 1729 reg_val = ((1ULL << q_idx) << NICVF_INTR_CQ_SHIFT); 1732 reg_val = ((1ULL << q_idx) << NICVF_INTR_SQ_SHIFT); 1735 reg_val = ((1ULL << q_idx) << NICVF_INTR_RBDR_SHIFT); 1757 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx) argument 1759 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1771 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx) argument 1773 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1785 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx) argument 1787 u64 mask = nicvf_int_type_to_mask(int_type, q_idx); 1799 nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx) argument [all...] |
H A D | nicvf_queues.h | 354 void nicvf_enable_intr(struct nicvf *nic, int int_type, int q_idx); 355 void nicvf_disable_intr(struct nicvf *nic, int int_type, int q_idx); 356 void nicvf_clear_intr(struct nicvf *nic, int int_type, int q_idx); 357 int nicvf_is_intr_enabled(struct nicvf *nic, int int_type, int q_idx);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | t3_hw.c | 2473 int i, j, cpu_idx = 0, q_idx = 0; local 2490 (i << 16) | rspq[q_idx++]); 2491 if (rspq[q_idx] == 0xffff) 2492 q_idx = 0;
|
/linux-master/drivers/net/ethernet/emulex/benet/ |
H A D | be_main.c | 1371 u16 q_idx = skb_get_queue_mapping(skb); local 1372 struct be_tx_obj *txo = &adapter->tx_obj[q_idx]; 1402 netif_stop_subqueue(netdev, q_idx); 1406 if (flush || __netif_subqueue_stopped(netdev, q_idx))
|
/linux-master/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_tx.c | 576 int i, q_idx; local 585 q_idx = smp_processor_id(); 586 if (unlikely(q_idx >= fp->num_xdpqs)) 589 for (q = xdpqs[q_idx], i = 0; i < n; i++)
|
/linux-master/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_main.c | 1874 int q_idx = pc; local 1877 interface->tx_ring[offset + i]->reg_idx = q_idx; 1879 interface->rx_ring[offset + i]->reg_idx = q_idx; 1881 q_idx += pc_stride;
|
H A D | fm10k_pci.c | 1175 int q_idx; local 1177 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { 1178 q_vector = interface->q_vector[q_idx]; 1872 int q_idx; local 1874 for (q_idx = 0; q_idx < interface->num_q_vectors; q_idx++) { 1875 q_vector = interface->q_vector[q_idx]; [all...] |
H A D | fm10k_pf.c | 503 u16 vsi, queue, pc, q_idx; local 520 q_idx = dglort->queue_b; 524 for (queue = 0; queue < queue_count; queue++, q_idx++) { 525 if (q_idx >= FM10K_MAX_QUEUES) 528 fm10k_write_reg(hw, FM10K_TX_SGLORT(q_idx), glort); 529 fm10k_write_reg(hw, FM10K_RX_SGLORT(q_idx), glort); 539 q_idx = pc + dglort->queue_b; 541 if (q_idx >= FM10K_MAX_QUEUES) 544 txqctl = fm10k_read_reg(hw, FM10K_TXQCTL(q_idx)); 547 fm10k_write_reg(hw, FM10K_TXQCTL(q_idx), txqct [all...] |