/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 39 static int iwch_build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, argument 49 wqe->send.rdmaop = T3_SEND_WITH_SE; 51 wqe->send.rdmaop = T3_SEND; 52 wqe->send.rem_stag = 0; 59 wqe->send.reserved[0] = 0; 60 wqe->send.reserved[1] = 0; 61 wqe->send.reserved[2] = 0; 64 wqe->send.sgl[0].stag = wr->imm_data; 65 wqe->send.sgl[0].len = __constant_cpu_to_be32(0); 66 wqe 88 iwch_build_rdma_write(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 129 iwch_build_rdma_read(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 197 iwch_build_rdma_recv(struct iwch_dev *rhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 243 union t3_wr *wqe; local 337 union t3_wr *wqe; local 389 union t3_wr *wqe; local 600 union t3_wr *wqe; local [all...] |
H A D | cxio_hal.c | 139 struct t3_modify_qp_wr *wqe; local 140 struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL); 145 wqe = (struct t3_modify_qp_wr *) skb_put(skb, sizeof(*wqe)); 146 memset(wqe, 0, sizeof(*wqe)); 147 build_fw_riwrh((struct fw_riwrh *) wqe, T3_WR_QP_MOD, 3, 1, qpid, 7); 148 wqe->flags = cpu_to_be32(MODQP_WRITE_EC); 150 wqe->sge_cmd = cpu_to_be64(sge_cmd); 503 struct t3_modify_qp_wr *wqe; local 588 __be64 *wqe; local 812 struct t3_rdma_init_wr *wqe; local [all...] |
H A D | cxio_dbg.c | 110 void cxio_dump_wqe(union t3_wr *wqe) argument 112 __be64 *data = (__be64 *)wqe;
|
H A D | cxio_wr.h | 349 static inline enum t3_wr_opcode fw_riwrh_opcode(struct fw_riwrh *wqe) argument 351 return G_FW_RIWR_OP(be32_to_cpu(wqe->op_seop_flags)); 354 static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op, argument 358 wqe->op_seop_flags = cpu_to_be32(V_FW_RIWR_OP(op) | 362 wqe->gen_tid_len = cpu_to_be32(V_FW_RIWR_GEN(genbit) | 366 ((union t3_wr *)wqe)->genbit.genbit = cpu_to_be64(genbit);
|
H A D | cxio_hal.h | 189 void cxio_dump_wqe(union t3_wr *wqe);
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/ |
H A D | ipath_uc.c | 40 static void complete_last_send(struct ipath_qp *qp, struct ipath_swqe *wqe, argument 46 (wqe->wr.send_flags & IB_SEND_SIGNALED)) { 47 wc->wr_id = wqe->wr.wr_id; 49 wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; 51 wc->byte_len = wqe->length; 61 wqe = get_swqe_ptr(qp, qp->s_last); 79 struct ipath_swqe *wqe; local 93 wqe = get_swqe_ptr(qp, qp->s_last); 101 complete_last_send(qp, wqe, &wc); 109 qp->s_psn = wqe [all...] |
H A D | ipath_rc.c | 40 static u32 restart_sge(struct ipath_sge_state *ss, struct ipath_swqe *wqe, argument 45 len = ((psn - wqe->psn) & IPATH_PSN_MASK) * pmtu; 46 ss->sge = wqe->sg_list[0]; 47 ss->sg_list = wqe->sg_list + 1; 48 ss->num_sge = wqe->wr.num_sge; 50 return wqe->length - len; 56 * @wqe: the work queue to initialize the QP's SGE from 60 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument 64 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, 216 struct ipath_swqe *wqe; local 666 struct ipath_swqe *wqe = get_swqe_ptr(qp, n); local 746 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local 813 struct ipath_swqe *wqe; local 1076 struct ipath_swqe *wqe; local [all...] |
H A D | ipath_ruc.c | 101 static int init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe) argument 108 for (i = j = 0; i < wqe->num_sge; i++) { 109 if (wqe->sg_list[i].length == 0) 112 if ((user && wqe->sg_list[i].lkey == 0) || 113 !ipath_lkey_ok(qp, &qp->r_sg_list[j], &wqe->sg_list[i], 116 qp->r_len += wqe->sg_list[i].length; 126 wc.wr_id = wqe->wr_id; 162 struct ipath_rwqe *wqe; local 189 wqe = get_rwqe_ptr(rq, tail); 192 } while (!wr_id_only && !init_sge(qp, wqe)); 245 struct ipath_swqe *wqe; local 513 struct ipath_swqe *wqe; local [all...] |
H A D | ipath_srq.c | 56 struct ipath_rwqe *wqe; local 78 wqe = get_rwqe_ptr(&srq->rq, wq->head); 79 wqe->wr_id = wr->wr_id; 80 wqe->num_sge = wr->num_sge; 82 wqe->sg_list[i] = wr->sg_list[i]; 286 struct ipath_rwqe *wqe; local 289 wqe = get_rwqe_ptr(&srq->rq, tail); 290 p->wr_id = wqe->wr_id; 291 p->num_sge = wqe->num_sge; 292 for (i = 0; i < wqe [all...] |
H A D | ipath_ud.c | 39 static int init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe, argument 47 for (i = j = 0; i < wqe->num_sge; i++) { 48 if (wqe->sg_list[i].length == 0) 51 if ((user && wqe->sg_list[i].lkey == 0) || 53 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 55 *lengthp += wqe->sg_list[i].length; 63 wc.wr_id = wqe->wr_id; 111 struct ipath_rwqe *wqe; local 178 wqe = get_rwqe_ptr(rq, tail); 181 if (init_sge(qp, wqe, [all...] |
H A D | ipath_qp.c | 418 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local 420 wc.wr_id = wqe->wr.wr_id; 421 wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; 995 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local 1014 wqe = get_swqe_ptr(qp, qp->s_last); 1015 wc->wr_id = wqe->wr.wr_id; 1016 wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode];
|
H A D | ipath_verbs.c | 295 struct ipath_rwqe *wqe; local 316 wqe = get_rwqe_ptr(&qp->r_rq, wq->head); 317 wqe->wr_id = wr->wr_id; 318 wqe->num_sge = wr->num_sge; 320 wqe->sg_list[i] = wr->sg_list[i];
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 1586 void *wqe; local 1614 wqe = get_send_wqe(qp, ind); 1616 qp->sq.last = wqe; 1618 ((struct mthca_next_seg *) wqe)->nda_op = 0; 1619 ((struct mthca_next_seg *) wqe)->ee_nds = 0; 1620 ((struct mthca_next_seg *) wqe)->flags = 1628 ((struct mthca_next_seg *) wqe)->imm = wr->imm_data; 1630 wqe += sizeof (struct mthca_next_seg); 1638 ((struct mthca_raddr_seg *) wqe)->raddr = 1640 ((struct mthca_raddr_seg *) wqe) 1826 void *wqe; local 1939 void *wqe; local 2209 void *wqe; local [all...] |
H A D | mthca_srq.c | 92 static inline int *wqe_to_link(void *wqe) argument 94 return (int *) (wqe + offsetof(struct mthca_next_seg, imm)); 149 void *wqe; local 174 wqe = get_wqe(srq, i); 176 *wqe_to_link(wqe) = i < srq->max - 1 ? i + 1 : -1; 178 for (scatter = wqe + sizeof (struct mthca_next_seg); 179 (void *) scatter < wqe + (1 << srq->wqe_shift); 498 void *wqe; local 515 wqe = get_wqe(srq, ind); 516 next_ind = *wqe_to_link(wqe); 623 void *wqe; local [all...] |
H A D | mthca_cq.c | 126 __be32 wqe; member in struct:mthca_cqe 140 __be32 wqe; member in struct:mthca_err_cqe 317 mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe)); 393 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe), 482 cqe->wqe = new_wqe; 515 be32_to_cpu(cqe->wqe)); 544 wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset) 550 u32 wqe = be32_to_cpu(cqe->wqe); local 552 wqe_index = wqe >> sr 556 s32 wqe; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 119 u32 *wqe = get_send_wqe(qp, n); local 123 wqe[i] = 0xffffffff; 1045 void *wqe) 1048 struct mlx4_wqe_mlx_seg *mlx = wqe; 1049 struct mlx4_wqe_inline_seg *inl = wqe + sizeof *mlx; 1190 void *wqe; local 1216 ctrl = wqe = get_send_wqe(qp, ind & (qp->sq.wqe_cnt - 1)); 1232 wqe += sizeof *ctrl; 1241 ((struct mlx4_wqe_raddr_seg *) wqe)->raddr = 1243 ((struct mlx4_wqe_raddr_seg *) wqe) 1044 build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe) argument [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ehca/ |
H A D | ehca_qp.c | 535 } else { /* for daqp we need to use msg size, not wqe size */ 687 * set purge bit of bad wqe and subsequent wqes to avoid reentering sqe 697 struct ehca_wqe *wqe; local 700 /* get send wqe pointer */ 713 /* convert wqe pointer to vadr */ 719 ehca_err(&shca->ib_device, "failed to get wqe offset qp_num=%x" 724 /* loop sets wqe's purge bit */ 725 wqe = (struct ehca_wqe*)ipz_qeit_calc(squeue, q_ofs); 727 while (wqe->optype != 0xff && wqe 896 struct ehca_wqe *wqe; local [all...] |