/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/bfa/ |
H A D | bfa_sgpg.c | 148 struct bfa_sgpg_wqe_s *wqe; local 164 wqe = bfa_q_first(&mod->sgpg_wait_q); 165 if (mod->free_sgpgs < wqe->nsgpg) 168 nsgpg = wqe->nsgpg; 169 bfa_sgpg_malloc(bfa, &wqe->sgpg_q, nsgpg); 170 wqe->nsgpg -= nsgpg; 171 if (wqe->nsgpg == 0) { 172 list_del(&wqe->qe); 173 wqe->cbfn(wqe 179 bfa_sgpg_wait(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe, int nsgpg) argument 205 bfa_sgpg_wcancel(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe) argument 218 bfa_sgpg_winit(struct bfa_sgpg_wqe_s *wqe, void (*cbfn) (void *cbarg), void *cbarg) argument [all...] |
H A D | bfa_sgpg_priv.h | 73 void bfa_sgpg_winit(struct bfa_sgpg_wqe_s *wqe, 75 void bfa_sgpg_wait(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe, 77 void bfa_sgpg_wcancel(struct bfa_s *bfa, struct bfa_sgpg_wqe_s *wqe);
|
H A D | bfa_intr_priv.h | 92 bfa_reqq_winit(struct bfa_reqq_wait_s *wqe, void (*qresume) (void *cbarg), argument 95 wqe->qresume = qresume; 96 wqe->cbarg = cbarg; 103 * bfa_reqq_wait(struct bfa_s *bfa, int reqq, struct bfa_reqq_wait_s *wqe)
|
H A D | bfa_fcxp.c | 204 struct bfa_fcxp_wqe_s *wqe; local 206 bfa_q_deq(&mod->wait_q, &wqe); 207 if (wqe) { 209 wqe->alloc_cbfn(wqe->alloc_cbarg, fcxp); 708 bfa_fcxp_alloc_wait(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe, argument 715 wqe->alloc_cbfn = alloc_cbfn; 716 wqe->alloc_cbarg = alloc_cbarg; 717 list_add_tail(&wqe->qe, &mod->wait_q); 721 bfa_fcxp_walloc_cancel(struct bfa_s *bfa, struct bfa_fcxp_wqe_s *wqe) argument [all...] |
H A D | bfa_intr.c | 41 struct bfa_reqq_wait_s *wqe; local 52 wqe = (struct bfa_reqq_wait_s *) qe; 53 wqe->qresume(wqe->cbarg);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 42 static int build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, argument 51 wqe->send.rdmaop = T3_SEND_WITH_SE; 53 wqe->send.rdmaop = T3_SEND; 54 wqe->send.rem_stag = 0; 58 wqe->send.rdmaop = T3_SEND_WITH_SE_INV; 60 wqe->send.rdmaop = T3_SEND_WITH_INV; 61 wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey); 68 wqe->send.reserved[0] = 0; 69 wqe->send.reserved[1] = 0; 70 wqe 87 build_rdma_write(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 128 build_rdma_read(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 149 build_fastreg(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt, int *wr_cnt, struct t3_wq *wq) argument 191 build_inv_stag(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 247 build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 287 build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 359 union t3_wr *wqe; local 471 union t3_wr *wqe; local 535 union t3_wr *wqe; local 743 union t3_wr *wqe; local 775 union t3_wr *wqe; local 862 union t3_wr *wqe = qhp->wq.queue; local [all...] |
H A D | cxio_hal.c | 140 struct t3_modify_qp_wr *wqe; local 141 struct sk_buff *skb = alloc_skb(sizeof(*wqe), GFP_KERNEL); 146 wqe = (struct t3_modify_qp_wr *) skb_put(skb, sizeof(*wqe)); 147 memset(wqe, 0, sizeof(*wqe)); 148 build_fw_riwrh((struct fw_riwrh *) wqe, T3_WR_QP_MOD, 151 wqe->flags = cpu_to_be32(MODQP_WRITE_EC); 153 wqe->sge_cmd = cpu_to_be64(sge_cmd); 516 struct t3_modify_qp_wr *wqe; local 601 __be64 *wqe; local 835 struct t3_rdma_init_wr *wqe; local [all...] |
H A D | cxio_dbg.c | 111 void cxio_dump_wqe(union t3_wr *wqe) argument 113 __be64 *data = (__be64 *)wqe;
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 114 int ipath_init_sge(struct ipath_qp *qp, struct ipath_rwqe *wqe, argument 121 for (i = j = 0; i < wqe->num_sge; i++) { 122 if (wqe->sg_list[i].length == 0) 126 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 128 *lengthp += wqe->sg_list[i].length; 137 wc.wr_id = wqe->wr_id; 163 struct ipath_rwqe *wqe; local 196 wqe = get_rwqe_ptr(rq, tail); 202 } while (!ipath_init_sge(qp, wqe, &qp->r_len, &qp->r_sge)); 203 qp->r_wr_id = wqe 255 struct ipath_swqe *wqe; local 690 ipath_send_complete(struct ipath_qp *qp, struct ipath_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | ipath_rc.c | 42 static u32 restart_sge(struct ipath_sge_state *ss, struct ipath_swqe *wqe, argument 47 len = ((psn - wqe->psn) & IPATH_PSN_MASK) * pmtu; 48 ss->sge = wqe->sg_list[0]; 49 ss->sg_list = wqe->sg_list + 1; 50 ss->num_sge = wqe->wr.num_sge; 52 return wqe->length - len; 58 * @wqe: the work queue to initialize the QP's SGE from 62 static void ipath_init_restart(struct ipath_qp *qp, struct ipath_swqe *wqe) argument 66 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, 218 struct ipath_swqe *wqe; local 727 struct ipath_swqe *wqe = get_swqe_ptr(qp, n); local 807 struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last); local 863 struct ipath_swqe *wqe; local 1125 struct ipath_swqe *wqe; local [all...] |
H A D | ipath_uc.c | 49 struct ipath_swqe *wqe; local 70 wqe = get_swqe_ptr(qp, qp->s_last); 71 ipath_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 84 wqe = get_swqe_ptr(qp, qp->s_cur); 97 qp->s_psn = wqe->psn = qp->s_next_psn; 98 qp->s_sge.sge = wqe->sg_list[0]; 99 qp->s_sge.sg_list = wqe->sg_list + 1; 100 qp->s_sge.num_sge = wqe->wr.num_sge; 101 qp->s_len = len = wqe->length; 102 switch (wqe [all...] |
H A D | ipath_ud.c | 61 struct ipath_rwqe *wqe; local 132 wqe = get_rwqe_ptr(rq, tail); 134 if (!ipath_init_sge(qp, wqe, &rlen, &rsge)) { 148 wc.wr_id = wqe->wr_id; 244 struct ipath_swqe *wqe; local 267 wqe = get_swqe_ptr(qp, qp->s_last); 268 ipath_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 275 wqe = get_swqe_ptr(qp, qp->s_cur); 281 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; 297 ipath_ud_loopback(qp, wqe); [all...] |
H A D | ipath_srq.c | 57 struct ipath_rwqe *wqe; local 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 80 wqe->wr_id = wr->wr_id; 81 wqe->num_sge = wr->num_sge; 83 wqe->sg_list[i] = wr->sg_list[i]; 281 struct ipath_rwqe *wqe; local 284 wqe = get_rwqe_ptr(&srq->rq, tail); 285 p->wr_id = wqe->wr_id; 286 p->num_sge = wqe->num_sge; 287 for (i = 0; i < wqe [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 304 static int build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, argument 316 wqe->send.sendop_pkd = cpu_to_be32( 319 wqe->send.sendop_pkd = cpu_to_be32( 321 wqe->send.stag_inv = 0; 325 wqe->send.sendop_pkd = cpu_to_be32( 328 wqe->send.sendop_pkd = cpu_to_be32( 330 wqe->send.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey); 340 ret = build_immd(sq, wqe->send.u.immd_src, wr, 344 size = sizeof wqe->send + sizeof(struct fw_ri_immd) + 349 wqe 369 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 412 build_rdma_read(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 441 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, struct ib_recv_wr *wr, u8 *len16) argument 456 build_fastreg(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 506 build_inv_stag(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 536 union t4_wr *wqe; local 647 union t4_recv_wr *wqe; local 847 union t4_wr *wqe; local 878 struct fw_ri_wr *wqe; local 967 struct fw_ri_wr *wqe; local 1041 struct fw_ri_wr *wqe; local [all...] |
H A D | t4.h | 107 static inline void init_wr_hdr(union t4_wr *wqe, u16 wrid, argument 110 wqe->send.opcode = (u8)opcode; 111 wqe->send.flags = flags; 112 wqe->send.wrid = wrid; 113 wqe->send.r1[0] = 0; 114 wqe->send.r1[1] = 0; 115 wqe->send.r1[2] = 0; 116 wqe->send.len16 = len16;
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/qib/ |
H A D | qib_rc.c | 43 static u32 restart_sge(struct qib_sge_state *ss, struct qib_swqe *wqe, argument 48 len = ((psn - wqe->psn) & QIB_PSN_MASK) * pmtu; 49 ss->sge = wqe->sg_list[0]; 50 ss->sg_list = wqe->sg_list + 1; 51 ss->num_sge = wqe->wr.num_sge; 52 ss->total_len = wqe->length; 54 return wqe->length - len; 237 struct qib_swqe *wqe; local 274 wqe = get_swqe_ptr(qp, qp->s_last); 276 qib_send_complete(qp, wqe, IB_WC_SUCCES 791 struct qib_swqe *wqe = get_swqe_ptr(qp, n); local 875 struct qib_swqe *wqe = get_swqe_ptr(qp, qp->s_acked); local 951 struct qib_swqe *wqe; local 977 struct qib_swqe *wqe; local 1058 do_rc_completion(struct qib_qp *qp, struct qib_swqe *wqe, struct qib_ibport *ibp) argument 1136 struct qib_swqe *wqe; local 1357 struct qib_swqe *wqe; local 1409 struct qib_swqe *wqe; local [all...] |
H A D | qib_ruc.c | 81 static int qib_init_sge(struct qib_qp *qp, struct qib_rwqe *wqe) argument 94 for (i = j = 0; i < wqe->num_sge; i++) { 95 if (wqe->sg_list[i].length == 0) 99 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 101 qp->r_len += wqe->sg_list[i].length; 117 wc.wr_id = wqe->wr_id; 144 struct qib_rwqe *wqe; local 176 wqe = get_rwqe_ptr(rq, tail); 185 if (!wr_id_only && !qib_init_sge(qp, wqe)) { 189 qp->r_wr_id = wqe 354 struct qib_swqe *wqe; local 768 qib_send_complete(struct qib_qp *qp, struct qib_swqe *wqe, enum ib_wc_status status) argument [all...] |
H A D | qib_uc.c | 49 struct qib_swqe *wqe; local 70 wqe = get_swqe_ptr(qp, qp->s_last); 71 qib_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 84 wqe = get_swqe_ptr(qp, qp->s_cur); 97 wqe->psn = qp->s_next_psn; 99 qp->s_sge.sge = wqe->sg_list[0]; 100 qp->s_sge.sg_list = wqe->sg_list + 1; 101 qp->s_sge.num_sge = wqe->wr.num_sge; 102 qp->s_sge.total_len = wqe->length; 103 len = wqe [all...] |
H A D | qib_ud.c | 238 struct qib_swqe *wqe; local 261 wqe = get_swqe_ptr(qp, qp->s_last); 262 qib_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); 269 wqe = get_swqe_ptr(qp, qp->s_cur); 277 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; 293 qib_ud_loopback(qp, wqe); 295 qib_send_complete(qp, wqe, IB_WC_SUCCESS); 301 extra_bytes = -wqe->length & 3; 302 nwords = (wqe->length + extra_bytes) >> 2; 306 qp->s_cur_size = wqe [all...] |
H A D | qib_srq.c | 57 struct qib_rwqe *wqe; local 79 wqe = get_rwqe_ptr(&srq->rq, wq->head); 80 wqe->wr_id = wr->wr_id; 81 wqe->num_sge = wr->num_sge; 83 wqe->sg_list[i] = wr->sg_list[i]; 277 struct qib_rwqe *wqe; local 280 wqe = get_rwqe_ptr(&srq->rq, tail); 281 p->wr_id = wqe->wr_id; 282 p->num_sge = wqe->num_sge; 283 for (i = 0; i < wqe [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mthca/ |
H A D | mthca_srq.c | 90 static inline int *wqe_to_link(void *wqe) argument 92 return (int *) (wqe + offsetof(struct mthca_next_seg, imm)); 147 void *wqe; local 174 next = wqe = get_wqe(srq, i); 177 *wqe_to_link(wqe) = i + 1; 180 *wqe_to_link(wqe) = -1; 184 for (scatter = wqe + sizeof (struct mthca_next_seg); 185 (void *) scatter < wqe + (1 << srq->wqe_shift); 502 void *wqe; local 511 wqe 601 void *wqe; local [all...] |
H A D | mthca_qp.c | 1616 void *wqe; local 1651 wqe = get_send_wqe(qp, ind); 1653 qp->sq.last = wqe; 1655 ((struct mthca_next_seg *) wqe)->nda_op = 0; 1656 ((struct mthca_next_seg *) wqe)->ee_nds = 0; 1657 ((struct mthca_next_seg *) wqe)->flags = 1665 ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; 1667 wqe += sizeof (struct mthca_next_seg); 1675 set_raddr_seg(wqe, wr->wr.atomic.remote_addr, 1677 wqe 1832 void *wqe; local 1929 void *wqe; local 2172 void *wqe; local [all...] |
H A D | mthca_cq.c | 126 __be32 wqe; member in struct:mthca_cqe 140 __be32 wqe; member in struct:mthca_err_cqe 312 mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe)); 388 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe), 477 cqe->wqe = new_wqe; 511 be32_to_cpu(cqe->wqe)); 540 wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset) 546 u32 wqe = be32_to_cpu(cqe->wqe); local 548 wqe_index = wqe >> sr 552 s32 wqe; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 135 __be32 *wqe; local 150 wqe = buf + (i & ((1 << qp->sq.wqe_shift) - 1)); 151 *wqe = stamp; 157 wqe = buf + i; 158 *wqe = cpu_to_be32(0xffffffff); 167 void *wqe; local 170 ctrl = wqe = get_send_wqe(qp, n & (qp->sq.wqe_cnt - 1)); 174 struct mlx4_wqe_datagram_seg *dgram = wqe + sizeof *ctrl; 183 inl = wqe + s; 1218 void *wqe, unsigne 1217 build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe, unsigned *mlx_seg_len) argument 1484 build_lso_seg(struct mlx4_wqe_lso_seg *wqe, struct ib_send_wr *wr, struct mlx4_ib_qp *qp, unsigned *lso_seg_len, __be32 *lso_hdr_sz, __be32 *blh) argument 1524 void *wqe; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.c | 215 struct nes_hw_qp_wqe *wqe; local 235 wqe = &nesqp->hwqp.sq_vbase[head]; 236 /* nes_debug(NES_DBG_MR, "processing sq wqe at %p, head = %u.\n", wqe, head); */ 237 nes_fill_init_qp_wqe(wqe, nesqp, head); 239 set_wqe_64bit_value(wqe->wqe_words, NES_IWARP_SQ_WQE_COMP_SCRATCH_LOW_IDX, u64temp); 254 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_WQE_MISC_IDX, wqe_misc); 255 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_BIND_WQE_MR_IDX, ibmw_bind->mr->lkey); 256 set_wqe_32bit_value(wqe->wqe_words, NES_IWARP_SQ_BIND_WQE_MW_IDX, ibmw->rkey); 257 set_wqe_32bit_value(wqe 3236 fill_wqe_sg_send(struct nes_hw_qp_wqe *wqe, struct ib_send_wr *ib_wr, u32 uselkey) argument 3270 struct nes_hw_qp_wqe *wqe; local 3548 struct nes_hw_qp_wqe *wqe; local [all...] |