Lines Matching refs:rx_oob
1990 struct mana_recv_buf_oob *rx_oob;
2018 rx_oob = &rxq->rx_oobs[i];
2020 if (!rx_oob->buf_va)
2023 dma_unmap_single(dev, rx_oob->sgl[0].address,
2024 rx_oob->sgl[0].size, DMA_FROM_DEVICE);
2026 page = virt_to_head_page(rx_oob->buf_va);
2028 if (rx_oob->from_pool)
2033 rx_oob->buf_va = NULL;
2044 static int mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key,
2060 rx_oob->buf_va = va;
2061 rx_oob->from_pool = from_pool;
2063 rx_oob->sgl[0].address = da;
2064 rx_oob->sgl[0].size = rxq->datasize;
2065 rx_oob->sgl[0].mem_key = mem_key;
2077 struct mana_recv_buf_oob *rx_oob;
2088 rx_oob = &rxq->rx_oobs[buf_idx];
2089 memset(rx_oob, 0, sizeof(*rx_oob));
2091 rx_oob->num_sge = 1;
2093 ret = mana_fill_rx_oob(rx_oob, apc->ac->gdma_dev->gpa_mkey, rxq,
2098 rx_oob->wqe_req.sgl = rx_oob->sgl;
2099 rx_oob->wqe_req.num_sge = rx_oob->num_sge;
2100 rx_oob->wqe_req.inline_oob_size = 0;
2101 rx_oob->wqe_req.inline_oob_data = NULL;
2102 rx_oob->wqe_req.flags = 0;
2103 rx_oob->wqe_req.client_data_unit = 0;
2106 MANA_WQE_SGE_SIZE * rx_oob->num_sge, 32);
2115 struct mana_recv_buf_oob *rx_oob;
2120 rx_oob = &rxq->rx_oobs[buf_idx];
2122 err = mana_gd_post_and_ring(rxq->gdma_rq, &rx_oob->wqe_req,
2123 &rx_oob->wqe_inf);