Lines Matching refs:mq
90 q = &cq->mq;
136 ce = (struct c2wr_ce *) c2_mq_consume(&cq->mq);
148 c2_mq_free(&cq->mq);
149 ce = (struct c2wr_ce *) c2_mq_consume(&cq->mq);
194 c2_mq_free(&cq->mq);
228 shared = cq->mq.peer;
248 ret = !c2_mq_empty(&cq->mq);
255 static void c2_free_cq_buf(struct c2_dev *c2dev, struct c2_mq *mq)
257 dma_free_coherent(&c2dev->pcidev->dev, mq->q_size * mq->msg_size,
258 mq->msg_pool.host, pci_unmap_addr(mq, mapping));
261 static int c2_alloc_cq_buf(struct c2_dev *c2dev, struct c2_mq *mq, int q_size,
267 &mq->host_dma, GFP_KERNEL);
271 c2_mq_rep_init(mq,
279 pci_unmap_addr_set(mq, mapping, mq->host_dma);
299 cq->mq.shared = c2_alloc_mqsp(c2dev, c2dev->kern_mqsp_pool,
300 &cq->mq.shared_dma, GFP_KERNEL);
301 if (!cq->mq.shared)
305 err = c2_alloc_cq_buf(c2dev, &cq->mq, entries + 1, C2_CQ_MSG_SIZE);
319 wr.msg_size = cpu_to_be32(cq->mq.msg_size);
320 wr.depth = cpu_to_be32(cq->mq.q_size);
321 wr.shared_ht = cpu_to_be64(cq->mq.shared_dma);
322 wr.msg_pool = cpu_to_be64(cq->mq.host_dma);
347 cq->mq.index = be32_to_cpu(reply->mq_index);
350 cq->mq.peer = ioremap_nocache(peer_pa, PAGE_SIZE);
351 if (!cq->mq.peer) {
367 cq->cqn = cq->mq.index;
377 c2_free_cq_buf(c2dev, &cq->mq);
379 c2_free_mqsp(cq->mq.shared);
395 c2dev->qptr_array[cq->mq.index] = NULL;
431 c2_free_cq_buf(c2dev, &cq->mq);