/linux-master/include/uapi/linux/ |
H A D | i2o-dev.h | 91 void __user *qbuf; /* Pointer to HTTP query string */ member in struct:i2o_html
|
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma.h | 22 void *qbuf; member in struct:erdma_eq 40 void *qbuf; member in struct:erdma_cmdq_sq 55 void *qbuf; member in struct:erdma_cmdq_cq 219 static inline void *get_queue_entry(void *qbuf, u32 idx, u32 depth, u32 shift) argument 223 return qbuf + (idx << shift);
|
H A D | erdma_cmdq.c | 99 sq->qbuf = 102 if (!sq->qbuf) 105 sq->db_record = (u64 *)(sq->qbuf + buf_size); 129 cq->qbuf = 132 if (!cq->qbuf) 137 cq->db_record = (u64 *)(cq->qbuf + buf_size); 160 eq->qbuf = 163 if (!eq->qbuf) 170 eq->db_record = (u64 *)(eq->qbuf + buf_size); 217 cmdq->cq.qbuf, cmd [all...] |
H A D | erdma_cq.c | 11 __be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci,
|
H A D | erdma_eq.c | 24 u64 *eqe = get_queue_entry(eq->qbuf, eq->ci, eq->depth, EQE_SHIFT); 91 eq->qbuf = 94 if (!eq->qbuf) 102 eq->db_record = (u64 *)(eq->qbuf + buf_size); 120 WARPPED_BUFSIZE(eq->depth << EQE_SHIFT), eq->qbuf, 235 eq->qbuf = 238 if (!eq->qbuf) 248 eq->db_record = (u64 *)(eq->qbuf + buf_size); 279 dma_free_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size), eq->qbuf,
|
H A D | erdma_verbs.c | 1242 cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr); 1602 cq->kern_cq.qbuf = 1606 if (!cq->kern_cq.qbuf) 1610 (u64 *)(cq->kern_cq.qbuf + (cq->depth << CQE_SHIFT)); 1681 cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
|
H A D | erdma_verbs.h | 241 void *qbuf; member in struct:erdma_kcq_info
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | t4vf_hw.c | 1300 u16 qbuf[3]; local 1301 u16 *qbp = qbuf; 1305 qbuf[0] = qbuf[1] = qbuf[2] = 0; 1312 *qp++ = cpu_to_be32(FW_RSS_IND_TBL_CMD_IQ0_V(qbuf[0]) | 1313 FW_RSS_IND_TBL_CMD_IQ1_V(qbuf[1]) | 1314 FW_RSS_IND_TBL_CMD_IQ2_V(qbuf[2]));
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_lag.c | 368 * @qbuf: pointer to buffer to populate 376 ice_lag_qbuf_recfg(struct ice_hw *hw, struct ice_aqc_cfg_txqs_buf *qbuf, argument 404 qbuf->queue_info[count].q_handle = cpu_to_le16(qid); 405 qbuf->queue_info[count].tc = tc; 406 qbuf->queue_info[count].q_teid = cpu_to_le32(q_ctx->q_teid); 498 struct ice_aqc_cfg_txqs_buf *qbuf; local 539 qbuf_size = struct_size(qbuf, queue_info, numq); 540 qbuf = kzalloc(qbuf_size, GFP_KERNEL); 541 if (!qbuf) { 547 valq = ice_lag_qbuf_recfg(&lag->pf->hw, qbuf, vsi_nu 856 struct ice_aqc_cfg_txqs_buf *qbuf; local 1880 struct ice_aqc_cfg_txqs_buf *qbuf; local [all...] |
/linux-master/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_ctrl_mbox.c | 135 u8 __iomem *qbuf; local 139 qbuf = (q->hw_q + *pi); 142 memcpy_toio(qbuf, buf, w_sz); 147 memcpy_toio(qbuf, buf, cp_sz); 153 qbuf = (q->hw_q + *pi); 154 memcpy_toio(qbuf, buf, w_sz); 200 u8 __iomem *qbuf; local 204 qbuf = (q->hw_q + *ci); 207 memcpy_fromio(buf, qbuf, r_sz); 212 memcpy_fromio(buf, qbuf, cp_s [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | dma.c | 645 struct mt76_queue_buf qbuf = {}; local 662 qbuf.addr = addr + q->buf_offset; 664 qbuf.len = len - q->buf_offset; 665 qbuf.skip_unmap = false; 666 if (mt76_dma_add_rx_buf(dev, q, &qbuf, buf) < 0) {
|