Searched refs:wqe (Results 1 - 17 of 17) sorted by relevance

/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/cxgb3/
H A Diwch_qp.c39 static int iwch_build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, argument
49 wqe->send.rdmaop = T3_SEND_WITH_SE;
51 wqe->send.rdmaop = T3_SEND;
52 wqe->send.rem_stag = 0;
59 wqe->send.reserved[0] = 0;
60 wqe->send.reserved[1] = 0;
61 wqe->send.reserved[2] = 0;
64 wqe->send.sgl[0].stag = wr->imm_data;
65 wqe->send.sgl[0].len = __constant_cpu_to_be32(0);
66 wqe
88 iwch_build_rdma_write(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument
129 iwch_build_rdma_read(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument
197 iwch_build_rdma_recv(struct iwch_dev *rhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument
243 union t3_wr *wqe; local
337 union t3_wr *wqe; local
389 union t3_wr *wqe; local
600 union t3_wr *wqe; local
[all...]
H A Dcxio_hal.c139 struct t3_modify_qp_wr *wqe; local
140 struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL);
145 wqe = (struct t3_modify_qp_wr *) skb_put(skb, sizeof(*wqe));
146 memset(wqe, 0, sizeof(*wqe));
147 build_fw_riwrh((struct fw_riwrh *) wqe, T3_WR_QP_MOD, 3, 1, qpid, 7);
148 wqe->flags = cpu_to_be32(MODQP_WRITE_EC);
150 wqe->sge_cmd = cpu_to_be64(sge_cmd);
503 struct t3_modify_qp_wr *wqe; local
588 __be64 *wqe; local
812 struct t3_rdma_init_wr *wqe; local
[all...]
H A Dcxio_dbg.c110 void cxio_dump_wqe(union t3_wr *wqe) argument
112 __be64 *data = (__be64 *)wqe;
H A Dcxio_wr.h349 static inline enum t3_wr_opcode fw_riwrh_opcode(struct fw_riwrh *wqe) argument
351 return G_FW_RIWR_OP(be32_to_cpu(wqe->op_seop_flags));
354 static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op, argument
358 wqe->op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(op) |
362 wqe->gen_tid_len = cpu_to_be32(V_FW_RIWR_GEN(genbit) |
366 ((union t3_wr *)wqe)->genbit.genbit = cpu_to_be64(genbit);
H A Dcxio_hal.h189 void cxio_dump_wqe(union t3_wr *wqe);
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/
H A Dipath_uc.c40 static void complete_last_send(struct ipath_qp *qp, struct ipath_swqe *wqe, argument
46 (wqe->wr.send_flags & IB_SEND_SIGNALED)) {
47 wc->wr_id = wqe->wr.wr_id;
49 wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode];
51 wc->byte_len = wqe->length;
61 wqe = get_swqe_ptr(qp, qp->s_last);
79 struct ipath_swqe *wqe; local
93 wqe = get_swqe_ptr(qp, qp->s_last);
101 complete_last_send(qp, wqe, &wc);
109 qp->s_psn = wqe
[all...]
H A Dipath_rc.c40 static u32 restart_sge(struct ipath_sge_state *ss, struct ipath_swqe *wqe, argument
45 len = ((psn - wqe->psn) & IPATH_PSN_MASK) * pmtu;
46 ss->sge = wqe->sg_list[0];
47 ss->sg_list = wqe->sg_list + 1;
48 ss->num_sge = wqe->wr.num_sge;
50 return wqe->length - len;
56 * @wqe: the work queue to initialize the QP's SGE from
60 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument
64 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn,
216 struct ipath_swqe *wqe; local
666 struct ipath_swqe *wqe = get_swqe_ptr(qp, n); local
746 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local
813 struct ipath_swqe *wqe; local
1076 struct ipath_swqe *wqe; local
[all...]
H A Dipath_ruc.c101 static int init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe) argument
108 for (i = j = 0; i < wqe->num_sge; i++) {
109 if (wqe->sg_list[i].length == 0)
112 if ((user && wqe->sg_list[i].lkey == 0) ||
113 !ipath_lkey_ok(qp, &qp->r_sg_list[j], &wqe->sg_list[i],
116 qp->r_len += wqe->sg_list[i].length;
126 wc.wr_id = wqe->wr_id;
162 struct ipath_rwqe *wqe; local
189 wqe = get_rwqe_ptr(rq, tail);
192 } while (!wr_id_only && !init_sge(qp, wqe));
245 struct ipath_swqe *wqe; local
513 struct ipath_swqe *wqe; local
[all...]
H A Dipath_srq.c56 struct ipath_rwqe *wqe; local
78 wqe = get_rwqe_ptr(&srq->rq, wq->head);
79 wqe->wr_id = wr->wr_id;
80 wqe->num_sge = wr->num_sge;
82 wqe->sg_list[i] = wr->sg_list[i];
286 struct ipath_rwqe *wqe; local
289 wqe = get_rwqe_ptr(&srq->rq, tail);
290 p->wr_id = wqe->wr_id;
291 p->num_sge = wqe->num_sge;
292 for (i = 0; i < wqe
[all...]
H A Dipath_ud.c39 static int init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe, argument
47 for (i = j = 0; i < wqe->num_sge; i++) {
48 if (wqe->sg_list[i].length == 0)
51 if ((user && wqe->sg_list[i].lkey == 0) ||
53 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE))
55 *lengthp += wqe->sg_list[i].length;
63 wc.wr_id = wqe->wr_id;
111 struct ipath_rwqe *wqe; local
178 wqe = get_rwqe_ptr(rq, tail);
181 if (init_sge(qp, wqe,
[all...]
H A Dipath_qp.c418 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local
420 wc.wr_id = wqe->wr.wr_id;
421 wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode];
995 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local
1014 wqe = get_swqe_ptr(qp, qp->s_last);
1015 wc->wr_id = wqe->wr.wr_id;
1016 wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode];
H A Dipath_verbs.c295 struct ipath_rwqe *wqe; local
316 wqe = get_rwqe_ptr(&qp->r_rq, wq->head);
317 wqe->wr_id = wr->wr_id;
318 wqe->num_sge = wr->num_sge;
320 wqe->sg_list[i] = wr->sg_list[i];
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c1586 void *wqe; local
1614 wqe = get_send_wqe(qp, ind);
1616 qp->sq.last = wqe;
1618 ((struct mthca_next_seg *) wqe)->nda_op = 0;
1619 ((struct mthca_next_seg *) wqe)->ee_nds = 0;
1620 ((struct mthca_next_seg *) wqe)->flags =
1628 ((struct mthca_next_seg *) wqe)->imm = wr->imm_data;
1630 wqe += sizeof (struct mthca_next_seg);
1638 ((struct mthca_raddr_seg *) wqe)->raddr =
1640 ((struct mthca_raddr_seg *) wqe)
1826 void *wqe; local
1939 void *wqe; local
2209 void *wqe; local
[all...]
H A Dmthca_srq.c92 static inline int *wqe_to_link(void *wqe) argument
94 return (int *) (wqe + offsetof(struct mthca_next_seg, imm));
149 void *wqe; local
174 wqe = get_wqe(srq, i);
176 *wqe_to_link(wqe) = i < srq->max - 1 ? i + 1 : -1;
178 for (scatter = wqe + sizeof (struct mthca_next_seg);
179 (void *) scatter < wqe + (1 << srq->wqe_shift);
498 void *wqe; local
515 wqe = get_wqe(srq, ind);
516 next_ind = *wqe_to_link(wqe);
623 void *wqe; local
[all...]
H A Dmthca_cq.c126 __be32 wqe; member in struct:mthca_cqe
140 __be32 wqe; member in struct:mthca_err_cqe
317 mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe));
393 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe),
482 cqe->wqe = new_wqe;
515 be32_to_cpu(cqe->wqe));
544 wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset)
550 u32 wqe = be32_to_cpu(cqe->wqe); local
552 wqe_index = wqe >> sr
556 s32 wqe; local
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mlx4/
H A Dqp.c119 u32 *wqe = get_send_wqe(qp, n); local
123 wqe[i] = 0xffffffff;
1045 void *wqe)
1048 struct mlx4_wqe_mlx_seg *mlx = wqe;
1049 struct mlx4_wqe_inline_seg *inl = wqe + sizeof *mlx;
1190 void *wqe; local
1216 ctrl = wqe = get_send_wqe(qp, ind & (qp->sq.wqe_cnt - 1));
1232 wqe += sizeof *ctrl;
1241 ((struct mlx4_wqe_raddr_seg *) wqe)->raddr =
1243 ((struct mlx4_wqe_raddr_seg *) wqe)
1044 build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe) argument
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ehca/
H A Dehca_qp.c535 } else { /* for daqp we need to use msg size, not wqe size */
687 * set purge bit of bad wqe and subsequent wqes to avoid reentering sqe
697 struct ehca_wqe *wqe; local
700 /* get send wqe pointer */
713 /* convert wqe pointer to vadr */
719 ehca_err(&shca->ib_device, "failed to get wqe offset qp_num=%x"
724 /* loop sets wqe's purge bit */
725 wqe = (struct ehca_wqe*)ipz_qeit_calc(squeue, q_ofs);
727 while (wqe->optype != 0xff && wqe
896 struct ehca_wqe *wqe; local
[all...]

Completed in 1032 milliseconds