Lines Matching refs:sqe
65 struct nvme_rdma_qe sqe;
290 kfree(req->sqe.data);
303 req->sqe.data = kzalloc(sizeof(struct nvme_command), GFP_KERNEL);
304 if (!req->sqe.data)
314 nvme_req(rq)->cmd = req->sqe.data;
1590 container_of(qe, struct nvme_rdma_request, sqe);
1675 struct nvme_rdma_qe *sqe = &ctrl->async_event_sqe;
1676 struct nvme_command *cmd = sqe->data;
1680 ib_dma_sync_single_for_cpu(dev, sqe->dma, sizeof(*cmd), DMA_TO_DEVICE);
1688 sqe->cqe.done = nvme_rdma_async_done;
1690 ib_dma_sync_single_for_device(dev, sqe->dma, sizeof(*cmd),
1693 ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL);
2004 struct nvme_rdma_qe *sqe = &req->sqe;
2018 req->sqe.dma = ib_dma_map_single(dev, req->sqe.data,
2021 err = ib_dma_mapping_error(dev, req->sqe.dma);
2025 ib_dma_sync_single_for_cpu(dev, sqe->dma,
2050 sqe->cqe.done = nvme_rdma_send_done;
2052 ib_dma_sync_single_for_device(dev, sqe->dma,
2055 err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge,
2073 ib_dma_unmap_single(dev, req->sqe.dma, sizeof(struct nvme_command),
2126 ib_dma_unmap_single(ibdev, req->sqe.dma, sizeof(struct nvme_command),