/linux-master/drivers/scsi/csiostor/ |
H A D | csio_scsi.c | 205 struct fw_scsi_cmd_wr *wr = (struct fw_scsi_cmd_wr *)addr; local 209 wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_CMD_WR) | 211 wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | 215 wr->cookie = (uintptr_t) req; 216 wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); 217 wr->tmo_val = (uint8_t) req->tmo; 218 wr->r3 = 0; 219 memset(&wr->r5, 0, 8); 225 wr->rsp_dmalen = cpu_to_be32(dma_buf->len); 226 wr 364 struct fw_scsi_read_wr *wr = (struct fw_scsi_read_wr *)wrp; local 417 struct fw_scsi_write_wr *wr = (struct fw_scsi_write_wr *)wrp; local 651 struct fw_scsi_abrt_cls_wr *wr = (struct fw_scsi_abrt_cls_wr *)addr; local 1102 csio_scsi_cmpl_handler(struct csio_hw *hw, void *wr, uint32_t len, struct csio_fl_dma_buf *flb, void *priv, uint8_t **scsiwr) argument [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/mei/ |
H A D | main.c | 310 u32 wr = le32_to_cpu(READ_ONCE(notif_q->wr_ptr)); local 314 if (rd > q_sz || wr > q_sz) { 320 room_in_buf = wr >= rd ? q_sz - wr + rd : rd - wr; 329 if (wr + tx_sz <= q_sz) { 330 memcpy(q_head + wr, hdr, tx_sz); 332 memcpy(q_head + wr, hdr, q_sz - wr); 333 memcpy(q_head, (const u8 *)hdr + q_sz - wr, tx_s 452 u32 wr; local 1007 iwl_mei_read_from_q(const u8 *q_head, u32 q_sz, u32 *_rd, u32 wr, void *_buf, u32 len) argument 1030 iwl_mei_handle_sap_data(struct mei_cl_device *cldev, const u8 *q_head, u32 q_sz, u32 rd, u32 wr, ssize_t valid_rx_sz, struct sk_buff_head *tx_skbs) argument 1104 iwl_mei_handle_sap_rx_cmd(struct mei_cl_device *cldev, const u8 *q_head, u32 q_sz, u32 rd, u32 wr, ssize_t valid_rx_sz) argument 1148 u32 wr = le32_to_cpu(READ_ONCE(notif_q->wr_ptr)); local [all...] |
/linux-master/drivers/infiniband/ulp/iser/ |
H A D | iser_verbs.c | 810 struct ib_recv_wr wr; local 818 wr.wr_cqe = &desc->cqe; 819 wr.sg_list = &desc->sge; 820 wr.num_sge = 1; 821 wr.next = NULL; 823 ret = ib_post_recv(ib_conn->qp, &wr, NULL); 833 struct ib_recv_wr wr; local 837 wr.wr_cqe = &rx_desc->cqe; 838 wr.sg_list = &rx_desc->rx_sg; 839 wr 859 struct ib_send_wr *wr = &tx_desc->send_wr; local [all...] |
/linux-master/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.c | 1875 int bnxt_re_post_srq_recv(struct ib_srq *ib_srq, const struct ib_recv_wr *wr, argument 1885 while (wr) { 1887 wqe.num_sge = wr->num_sge; 1888 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); 1889 wqe.wr_id = wr->wr_id; 1894 *bad_wr = wr; 1897 wr = wr->next; 2242 const struct ib_send_wr *wr, 2241 bnxt_re_build_qp1_send_v2(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument 2415 bnxt_re_build_qp1_shadow_qp_recv(struct bnxt_re_qp *qp, const struct ib_recv_wr *wr, struct bnxt_qplib_swqe *wqe, int payload_size) argument 2464 bnxt_re_build_send_wqe(struct bnxt_re_qp *qp, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2503 bnxt_re_build_rdma_wqe(const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2535 bnxt_re_build_atomic_wqe(const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2562 bnxt_re_build_inv_wqe(const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2576 bnxt_re_build_reg_wqe(const struct ib_reg_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2612 bnxt_re_copy_inline_data(struct bnxt_re_dev *rdev, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2642 bnxt_re_copy_wr_payload(struct bnxt_re_dev *rdev, const struct ib_send_wr *wr, struct bnxt_qplib_swqe *wqe) argument 2673 bnxt_re_post_send_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_send_wr *wr) argument 2732 bnxt_re_post_send(struct ib_qp *ib_qp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 2833 bnxt_re_post_recv_shadow_qp(struct bnxt_re_dev *rdev, struct bnxt_re_qp *qp, const struct ib_recv_wr *wr) argument 2867 bnxt_re_post_recv(struct ib_qp *ib_qp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/lib/ |
H A D | test_bitmap.c | 809 unsigned int wr, bit; local 820 for (wr = 0; wr < 500; wr++) { 823 for_each_set_bit_wrap(bit, orig, 500, wr) 856 unsigned int wr, bit; local 867 for (wr = 0; wr < 500; wr++) { 871 bit = wr; 908 unsigned int wr, bit; local 982 unsigned int wr, s, e; local 1012 unsigned int wr, s, e; local [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 625 for (i = 0; i < wqe->wr.num_sge; i++) { 1774 * @wr: the WR to post 1781 int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, argument 1792 *bad_wr = wr; 1796 for (; wr; wr = wr->next) { 1801 if ((unsigned)wr->num_sge > qp->r_rq.max_sge) { 1802 *bad_wr = wr; 1812 *bad_wr = wr; 1862 rvt_qp_valid_operation( struct rvt_qp *qp, const struct rvt_operation_params *post_parms, const struct ib_send_wr *wr) argument 1960 rvt_post_one_wr(struct rvt_qp *qp, const struct ib_send_wr *wr, bool *call_send) argument 2143 rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 2204 rvt_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib_uc.c | 101 qp->s_sge.num_sge = wqe->wr.num_sge; 105 switch (wqe->wr.opcode) { 113 if (wqe->wr.opcode == IB_WR_SEND) 119 ohdr->u.imm_data = wqe->wr.ex.imm_data; 122 if (wqe->wr.send_flags & IB_SEND_SOLICITED) 142 if (wqe->wr.opcode == IB_WR_RDMA_WRITE) 148 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; 150 if (wqe->wr.send_flags & IB_SEND_SOLICITED) 172 if (wqe->wr.opcode == IB_WR_SEND) 177 ohdr->u.imm_data = wqe->wr [all...] |
/linux-master/net/9p/ |
H A D | trans_rdma.c | 386 struct ib_recv_wr wr; local 402 wr.next = NULL; 403 wr.wr_cqe = &c->cqe; 404 wr.sg_list = &sge; 405 wr.num_sge = 1; 407 ret = ib_post_recv(rdma->qp, &wr, NULL); 421 struct ib_send_wr wr; local 499 wr.next = NULL; 500 wr.wr_cqe = &c->cqe; 501 wr [all...] |
/linux-master/drivers/infiniband/hw/erdma/ |
H A D | erdma_qp.c | 330 rdma_wr = container_of(send_wr, struct ib_rdma_wr, wr); 357 rdma_wr = container_of(send_wr, struct ib_rdma_wr, wr); 504 const struct ib_send_wr *wr = send_wr; local 514 while (wr) { 521 ret = erdma_push_one_sqe(qp, &sq_pi, wr); 523 *bad_send_wr = wr; 529 wr = wr->next; 573 const struct ib_recv_wr *wr = recv_wr; local 580 while (wr) { [all...] |
/linux-master/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 1874 const struct ib_send_wr *wr) 1878 struct ocrdma_ah *ah = get_ocrdma_ah(ud_wr(wr)->ah); 1880 ud_hdr->rsvd_dest_qpn = ud_wr(wr)->remote_qpn; 1884 ud_hdr->qkey = ud_wr(wr)->remote_qkey; 1921 const struct ib_send_wr *wr, u32 wqe_size) 1926 if (wr->send_flags & IB_SEND_INLINE && qp->qp_type != IB_QPT_UD) { 1927 hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); 1935 for (i = 0; i < wr->num_sge; i++) { 1937 (void *)(unsigned long)wr 1872 ocrdma_build_ud_hdr(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1918 ocrdma_build_inline_sges(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, struct ocrdma_sge *sge, const struct ib_send_wr *wr, u32 wqe_size) argument 1958 ocrdma_build_send(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1975 ocrdma_build_write(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 1993 ocrdma_build_read(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_send_wr *wr) argument 2022 ocrdma_build_reg(struct ocrdma_qp *qp, struct ocrdma_hdr_wqe *hdr, const struct ib_reg_wr *wr) argument 2085 ocrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 2197 ocrdma_build_rqe(struct ocrdma_hdr_wqe *rqe, const struct ib_recv_wr *wr, u16 tag) argument 2218 ocrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 2288 ocrdma_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | srq.c | 402 int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, argument 419 *bad_wr = wr; 423 for (nreq = 0; wr; nreq++, wr = wr->next) { 424 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { 426 *bad_wr = wr; 432 *bad_wr = wr; 436 srq->wrid[srq->head] = wr->wr_id; 442 for (i = 0; i < wr [all...] |
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | rc.c | 495 if ((wqe->wr.send_flags & IB_SEND_FENCE) && 497 (wqe->wr.opcode != IB_WR_TID_RDMA_READ || 506 if (wqe->wr.opcode == IB_WR_REG_MR || 507 wqe->wr.opcode == IB_WR_LOCAL_INV) { 517 if (!(wqe->wr.send_flags & 521 wqe->wr.ex.invalidate_rkey); 552 switch (wqe->wr.opcode) { 564 if (wqe->wr.opcode == IB_WR_SEND) { 566 } else if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) { 569 ohdr->u.imm_data = wqe->wr [all...] |
/linux-master/drivers/media/radio/ |
H A D | radio-tea5764.c | 162 struct tea5764_write_regs wr; local 167 .len = sizeof(wr), 168 .buf = (void *)&wr 171 wr.intreg = r->intreg & 0xff; 172 wr.frqset = __cpu_to_be16(r->frqset); 173 wr.tnctrl = __cpu_to_be16(r->tnctrl); 174 wr.testreg = __cpu_to_be16(r->testreg); 175 wr.rdsctrl = __cpu_to_be16(r->rdsctrl); 176 wr.rdsbbl = __cpu_to_be16(r->rdsbbl);
|
/linux-master/drivers/infiniband/hw/qedr/ |
H A D | verbs.c | 3291 const struct ib_send_wr *wr, 3295 u32 data_size = sge_data_len(wr->sg_list, wr->num_sge); 3301 *bad_wr = wr; 3315 for (i = 0; i < wr->num_sge; i++) { 3316 u32 len = wr->sg_list[i].length; 3317 void *src = (void *)(uintptr_t)wr->sg_list[i].addr; 3376 const struct ib_send_wr *wr) 3381 for (i = 0; i < wr->num_sge; i++) { 3384 DMA_REGPAIR_LE(sge->addr, wr 3289 qedr_prepare_sq_inline_data(struct qedr_dev *dev, struct qedr_qp *qp, u8 *wqe_size, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr, u8 *bits, u8 bit) argument 3375 qedr_prepare_sq_sges(struct qedr_qp *qp, u8 *wqe_size, const struct ib_send_wr *wr) argument 3396 qedr_prepare_sq_rdma_data(struct qedr_dev *dev, struct qedr_qp *qp, struct rdma_sq_rdma_wqe_1st *rwqe, struct rdma_sq_rdma_wqe_2nd *rwqe2, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 3419 qedr_prepare_sq_send_data(struct qedr_dev *dev, struct qedr_qp *qp, struct rdma_sq_send_wqe_1st *swqe, struct rdma_sq_send_wqe_2st *swqe2, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 3438 qedr_prepare_reg(struct qedr_qp *qp, struct rdma_sq_fmr_wqe_1st *fwqe1, const struct ib_reg_wr *wr) argument 3502 qedr_can_post_send(struct qedr_qp *qp, const struct ib_send_wr *wr) argument 3540 __qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 3755 qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 3831 qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 3912 qedr_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 397 mr->mr_regwr.wr.next = post_wr; 398 mr->mr_regwr.wr.wr_cqe = &mr->mr_cqe; 399 mr->mr_regwr.wr.num_sge = 0; 400 mr->mr_regwr.wr.opcode = IB_WR_REG_MR; 401 mr->mr_regwr.wr.send_flags = 0; 402 post_wr = &mr->mr_regwr.wr; 689 mr->mr_regwr.wr.next = NULL; 690 mr->mr_regwr.wr.wr_cqe = &mr->mr_cqe; 691 mr->mr_regwr.wr.num_sge = 0; 692 mr->mr_regwr.wr [all...] |
/linux-master/arch/sparc/lib/ |
H A D | xor.S | 34 0: wr %g0, FPRS_FEF, %fprs 36 wr %g0, ASI_BLK_P, %asi 91 wr %g1, %g0, %asi 93 wr %g0, 0, %fprs 104 0: wr %g0, FPRS_FEF, %fprs 106 wr %g0, ASI_BLK_P, %asi 158 wr %g1, %g0, %asi 160 wr %g0, 0, %fprs 171 0: wr %g0, FPRS_FEF, %fprs 173 wr [all...] |
H A D | NGbzero.S | 45 wr %o4, 0x0, %asi 72 wr %g7, 0x0, %asi 89 wr %o4, 0x0, %asi 111 wr %o5, 0x0, %asi
|
/linux-master/arch/mips/kernel/ |
H A D | r4k_fpu.S | 295 .macro save_msa_upper wr, off, base 299 copy_s_d \wr, 1 302 copy_s_w \wr, 2 304 copy_s_w \wr, 3 307 copy_s_w \wr, 2 309 copy_s_w \wr, 3 352 .macro restore_msa_upper wr, off, base 357 insert_d \wr, 1 360 insert_w \wr, 2 362 insert_w \wr, [all...] |
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.h | 60 int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr, 62 int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr, 81 int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr,
|
/linux-master/arch/sparc/kernel/ |
H A D | wof.S | 111 wr %glob_tmp, 0x0, %wim ! set new %wim, this is safe now 122 wr %t_psr, 0x0, %psr ! restore condition codes in %psr 148 wr %glob_tmp, 0x0, %wim ! Now it is safe to set new %wim 191 wr %t_psr, 0x0, %psr 253 wr %t_psr, PSR_ET, %psr 286 wr %t_psr, 0x0, %psr
|
/linux-master/fs/autofs/ |
H A D | waitq.c | 51 ssize_t wr = 0; local 57 wr = __kernel_write(file, data, bytes, NULL); 58 if (wr <= 0) 60 data += wr; 61 bytes -= wr; 68 if (wr == -EPIPE && !sigpipe) { 75 /* if 'wr' returned 0 (impossible) we assume -EIO (safe) */ 76 return bytes == 0 ? 0 : wr < 0 ? wr : -EIO;
|
/linux-master/tools/testing/selftests/net/ |
H A D | tcp_mmap.c | 592 int64_t wr = FILE_SZ - total; local 594 if (wr > chunk_size - offset) 595 wr = chunk_size - offset; 597 wr = send(fd, buffer + offset, 598 (size_t)wr, zflg ? MSG_ZEROCOPY : 0); 599 if (wr <= 0) 602 EVP_DigestUpdate(ctx, buffer + offset, wr); 603 total += wr;
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_hw_v2.c | 104 * The hns wr opcode real value is start from 0, In order to distinguish between 136 const struct ib_reg_wr *wr) 140 struct hns_roce_mr *mr = to_hr_mr(wr->mr); 144 hr_reg_write_bool(fseg, FRMR_BIND_EN, wr->access & IB_ACCESS_MW_BIND); 146 wr->access & IB_ACCESS_REMOTE_ATOMIC); 147 hr_reg_write_bool(fseg, FRMR_RR, wr->access & IB_ACCESS_REMOTE_READ); 148 hr_reg_write_bool(fseg, FRMR_RW, wr->access & IB_ACCESS_REMOTE_WRITE); 149 hr_reg_write_bool(fseg, FRMR_LW, wr->access & IB_ACCESS_LOCAL_WRITE); 156 rc_sq_wqe->byte_16 = cpu_to_le32(wr->mr->length & 0xffffffff); 157 rc_sq_wqe->byte_20 = cpu_to_le32(wr 135 set_frmr_seg(struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, const struct ib_reg_wr *wr) argument 167 set_atomic_seg(const struct ib_send_wr *wr, struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, unsigned int valid_num_sge) argument 190 fill_ext_sge_inl_data(struct hns_roce_qp *qp, const struct ib_send_wr *wr, unsigned int *sge_idx, u32 msg_len) argument 283 set_rc_inl(struct hns_roce_qp *qp, const struct ib_send_wr *wr, struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, unsigned int *sge_idx) argument 328 set_rwqe_data_seg(struct ib_qp *ibqp, const struct ib_send_wr *wr, struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, unsigned int *sge_ind, unsigned int valid_num_sge) argument 392 calc_wr_sge_num(const struct ib_send_wr *wr, unsigned int *sge_len) argument 410 get_immtdata(const struct ib_send_wr *wr) argument 421 set_ud_opcode(struct hns_roce_v2_ud_send_wqe *ud_sq_wqe, const struct ib_send_wr *wr) argument 462 set_ud_wqe(struct hns_roce_qp *qp, const struct ib_send_wr *wr, void *wqe, unsigned int *sge_idx, unsigned int owner_bit) argument 519 set_rc_opcode(struct hns_roce_dev *hr_dev, struct hns_roce_v2_rc_send_wqe *rc_sq_wqe, const struct ib_send_wr *wr) argument 564 set_rc_wqe(struct hns_roce_qp *qp, const struct ib_send_wr *wr, void *wqe, unsigned int *sge_idx, unsigned int owner_bit) argument 684 hns_roce_v2_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, const struct ib_send_wr **bad_wr) argument 773 fill_recv_sge_to_wqe(const struct ib_recv_wr *wr, void *wqe, u32 max_sge, bool rsv) argument 800 fill_rq_wqe(struct hns_roce_qp *hr_qp, const struct ib_recv_wr *wr, u32 wqe_idx, u32 max_sge) argument 809 hns_roce_v2_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument 891 check_post_srq_valid(struct hns_roce_srq *srq, u32 max_sge, const struct ib_recv_wr *wr) argument 952 hns_roce_v2_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, const struct ib_recv_wr **bad_wr) argument [all...] |
/linux-master/net/smc/ |
H A D | smc_tx.c | 324 rdma_wr->wr.wr_id = smc_wr_tx_get_next_wr_id(link); 325 rdma_wr->wr.num_sge = num_sges; 333 rc = ib_post_send(link->roce_qp, &rdma_wr->wr, NULL); 373 struct ib_rdma_wr *wr = &wr_rdma_buf->wr_tx_rdma[dstchunk]; local 374 struct ib_sge *sge = wr->wr.sg_list; 379 wr->wr.send_flags |= IB_SEND_INLINE; 381 wr->wr [all...] |
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 1055 u64 *wr = (u64 *)&q->desc[index]; local 1059 wr); 1461 static void *write_eo_udp_wr(struct sk_buff *skb, struct fw_eth_tx_eo_wr *wr, argument 1464 wr->u.udpseg.type = FW_ETH_TX_EO_TYPE_UDPSEG; 1465 wr->u.udpseg.ethlen = skb_network_offset(skb); 1466 wr->u.udpseg.iplen = cpu_to_be16(skb_network_header_len(skb)); 1467 wr->u.udpseg.udplen = sizeof(struct udphdr); 1468 wr->u.udpseg.rtplen = 0; 1469 wr->u.udpseg.r4 = 0; 1471 wr 1497 struct fw_eth_tx_pkt_wr *wr; local 1814 struct fw_eth_tx_pkt_vm_wr *wr; local 2186 write_eo_wr(struct adapter *adap, struct sge_eosw_txq *eosw_txq, struct sk_buff *skb, struct fw_eth_tx_eo_wr *wr, u32 hdr_len, u32 wrlen) argument 2258 struct fw_eth_tx_eo_wr *wr; local 2639 ctrlq_check_stop(struct sge_ctrl_txq *q, struct fw_wr_hdr *wr) argument 2657 struct fw_eth_tx_pkt_wr *wr; local 2738 struct fw_wr_hdr *wr; local 2787 struct fw_wr_hdr *wr; local 2909 ofldtxq_stop(struct sge_uld_txq *q, struct fw_wr_hdr *wr) argument [all...] |