/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_cq.c | 40 * 1 cqe returned 48 struct t3_cqe cqe, *rd_cqe; local 67 ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, 83 wc->vendor_err = CQE_STATUS(cqe); 88 CQE_QPID(cqe), CQE_TYPE(cqe), 89 CQE_OPCODE(cqe), CQE_STATUS(cqe), CQE_WRID_HI(cqe), 90 CQE_WRID_LOW(cqe), (unsigne [all...] |
H A D | iwch_ev.c | 51 qhp = get_qhp(rnicp, CQE_QPID(rsp_msg->cqe)); 55 __func__, CQE_STATUS(rsp_msg->cqe), 56 CQE_QPID(rsp_msg->cqe)); 65 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 72 CQE_QPID(rsp_msg->cqe), CQE_OPCODE(rsp_msg->cqe), 73 CQE_STATUS(rsp_msg->cqe), CQE_TYPE(rsp_msg->cqe), 74 CQE_WRID_HI(rsp_msg->cqe), CQE_WRID_LOW(rsp_msg->cqe)); [all...] |
H A D | cxio_hal.c | 75 struct t3_cqe *cqe; local 105 * Now rptr is the index for the (last) cqe that was 109 cqe = cq->queue + Q_PTR2IDX(rptr, cq->size_log2); 110 while (!CQ_VLD_ENTRY(rptr, cq->size_log2, cqe)) { 349 struct t3_cqe cqe; local 353 memset(&cqe, 0, sizeof(cqe)); 354 cqe.header = cpu_to_be32(V_CQE_STATUS(TPT_ERR_SWFLUSH) | 361 *(cq->sw_queue + Q_PTR2IDX(cq->sw_wptr, cq->size_log2)) = cqe; 386 struct t3_cqe cqe; local 427 struct t3_cqe *cqe, *swcqe; local 443 cqe_completes_wr(struct t3_cqe *cqe, struct t3_wq *wq) argument 463 struct t3_cqe *cqe; local 481 struct t3_cqe *cqe; local 1146 cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument [all...] |
H A D | cxio_wr.h | 675 struct t3_cqe cqe; member in struct:t3_swsq 728 #define CQ_VLD_ENTRY(ptr,size_log2,cqe) (Q_GENBIT(ptr,size_log2) == \ 729 CQE_GENBIT(*cqe)) 753 struct t3_cqe *cqe; local 755 cqe = cq->queue + (Q_PTR2IDX(cq->rptr, cq->size_log2)); 756 if (CQ_VLD_ENTRY(cq->rptr, cq->size_log2, cqe)) 757 return cqe; 763 struct t3_cqe *cqe; local 766 cqe = cq->sw_queue + (Q_PTR2IDX(cq->sw_rptr, cq->size_log2)); 767 return cqe; 774 struct t3_cqe *cqe; local [all...] |
H A D | cxio_hal.h | 146 struct t3_cqe cqe; /* flits 2-3 */ member in struct:respQ_msg_t 195 int cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe,
|
H A D | iwch_provider.c | 196 chp->ibcq.cqe = 1 << chp->cq.size_log2; 237 static int iwch_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata) argument 244 PDBG("%s ib_cq %p cqe %d\n", __func__, cq, cqe); 247 if (cqe <= cq->cqe) 251 cqe = roundup_pow_of_two(cqe+1); 252 newcq.size_log2 = ilog2(cqe); 255 if (cqe < Q_COUN [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 174 static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe) argument 176 return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe; 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); 184 static inline void set_cqe_hw(struct mthca_cqe *cqe) argument 186 cqe->owner = MTHCA_CQ_ENTRY_OWNER_HW; 191 __be32 *cqe = cqe_ptr; local 193 (void) cqe; /* avoid warning if mthca_dbg compiled away... */ 195 be32_to_cpu(cqe[0]), be32_to_cpu(cqe[ 269 is_recv_cqe(struct mthca_cqe *cqe) argument 281 struct mthca_cqe *cqe; local 316 cqe, MTHCA_CQ_ENTRY_SIZE); local 371 mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe) argument 377 handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp *qp, int wqe_index, int is_send, struct mthca_err_cqe *cqe, struct ib_wc *entry, int *free_cqe) argument 490 struct mthca_cqe *cqe; local [all...] |
H A D | mthca_provider.h | 193 int cqe; member in struct:mthca_cq_resize
|
H A D | mthca_provider.c | 788 cq->resize_buf->cqe = entries - 1; 812 if (entries == ibcq->cqe + 1) { 837 cq->resize_buf->cqe); 854 tcqe = cq->ibcq.cqe; 856 cq->ibcq.cqe = cq->resize_buf->cqe; 859 tcqe = cq->resize_buf->cqe; 868 ibcq->cqe = entries - 1;
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 79 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); local 81 return (!!(cqe->owner_sr_opcode & MLX4_CQE_OWNER_MASK) ^ 82 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; 130 static void mlx4_ib_free_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int cqe) argument 132 mlx4_buf_free(dev->dev, (cqe + 1) * sizeof(struct mlx4_cqe), &buf->buf); 137 u64 buf_addr, int cqe) 141 *umem = ib_umem_get(context, buf_addr, cqe * sizeof (struct mlx4_cqe), 183 cq->ibcq.cqe 135 mlx4_ib_get_cq_umem(struct mlx4_ib_dev *dev, struct ib_ucontext *context, struct mlx4_ib_cq_buf *buf, struct ib_umem **umem, u64 buf_addr, int cqe) argument 329 struct mlx4_cqe *cqe, *new_cqe; local 455 dump_cqe(void *cqe) argument 465 mlx4_ib_handle_error_cqe(struct mlx4_err_cqe *cqe, struct ib_wc *wc) argument 544 struct mlx4_cqe *cqe; local 765 struct mlx4_cqe *cqe, *dest; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 193 struct t4_cqe cqe; local 197 memset(&cqe, 0, sizeof(cqe)); 198 cqe.header = cpu_to_be32(V_CQE_STATUS(T4_ERR_SWFLUSH) | 203 cqe.bits_type_ts = cpu_to_be64(V_CQE_GENBIT((u64)cq->gen)); 204 cq->sw_queue[cq->sw_pidx] = cqe; 226 struct t4_cqe cqe; local 230 memset(&cqe, 0, sizeof(cqe)); 231 cqe 265 struct t4_cqe *cqe = NULL, *swcqe; local 282 cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq) argument 300 struct t4_cqe *cqe; local 319 struct t4_cqe *cqe; local 416 poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument 577 struct t4_cqe cqe = {0, 0}, *rd_cqe; local 882 c4iw_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata) argument [all...] |
H A D | t4.h | 184 /* macros for flit 0 of the cqe */ 238 /* macros for flit 3 of the cqe */ 261 struct t4_cqe cqe; member in struct:t4_swsqe 488 static inline int t4_valid_cqe(struct t4_cq *cq, struct t4_cqe *cqe) argument 490 return (CQE_GENBIT(cqe) == cq->gen); 493 static inline int t4_next_hw_cqe(struct t4_cq *cq, struct t4_cqe **cqe) argument 508 *cqe = &cq->queue[cq->cidx]; 522 static inline int t4_next_cqe(struct t4_cq *cq, struct t4_cqe **cqe) argument 529 *cqe = &cq->sw_queue[cq->sw_cidx]; 531 ret = t4_next_hw_cqe(cq, cqe); [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/ipath/ |
H A D | ipath_cq.c | 63 if (head >= (unsigned) cq->ibcq.cqe) { 64 head = cq->ibcq.cqe; 147 if (tail > (u32) cq->ibcq.cqe) 148 tail = (u32) cq->ibcq.cqe; 154 if (tail >= cq->ibcq.cqe) 279 * ib_create_cq() will initialize cq->ibcq except for cq->ibcq.cqe. 283 cq->ibcq.cqe = entries; 371 int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument 380 if (cqe < 1 || cqe > ib_ipath_max_cqe [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/qib/ |
H A D | qib_cq.c | 63 if (head >= (unsigned) cq->ibcq.cqe) { 64 head = cq->ibcq.cqe; 145 if (tail > (u32) cq->ibcq.cqe) 146 tail = (u32) cq->ibcq.cqe; 152 if (tail >= cq->ibcq.cqe) 285 * ib_create_cq() will initialize cq->ibcq except for cq->ibcq.cqe. 289 cq->ibcq.cqe = entries; 377 int qib_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) argument 386 if (cqe < 1 || cqe > ib_qib_max_cqe [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/ehca/ |
H A D | ehca_reqs.c | 630 struct ehca_cqe *cqe; local 637 cqe = (struct ehca_cqe *) 639 if (!cqe) { 651 if (unlikely(cqe->status & WC_STATUS_PURGE_BIT)) { 656 qp = ehca_cq_get_qp(my_cq, cqe->local_qp_number); 659 "could not find qp -> ignore cqe", 660 my_cq->cq_number, cqe->local_qp_number); 661 ehca_dmp(cqe, 64, "cq_num=%x qp_num=%x", 662 my_cq->cq_number, cqe->local_qp_number); 663 /* ignore this purged cqe */ [all...] |
H A D | ehca_cq.c | 116 struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, int comp_vector, argument 134 if (cqe >= 0xFFFFFFFF - 64 - additional_cqe) 196 param.nr_cqe = cqe + additional_cqe; 274 my_cq->ib_cq.cqe = my_cq->nr_of_entries = 381 int ehca_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata) argument
|
H A D | ipz_pt_fn.h | 144 struct ehca_cqe *cqe = ipz_qeit_get(queue); local 145 return ((cqe->cqe_flags >> 7) == (queue->toggle_state & 1));
|
H A D | ehca_iverbs.h | 129 struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, int comp_vector, 135 int ehca_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/mlx4/ |
H A D | en_rx.c | 548 struct mlx4_cqe *cqe; local 565 * reading 'cqe->index' */ 567 cqe = &cq->buf[index]; 570 while (XNOR(cqe->owner_sr_opcode & MLX4_CQE_OWNER_MASK, 582 if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) == 586 ((struct mlx4_err_cqe *) cqe)->vendor_err_syndrome, 587 ((struct mlx4_err_cqe *) cqe)->syndrome); 590 if (unlikely(cqe->badfcs_enc & MLX4_CQE_BAD_FCS)) { 598 length = be32_to_cpu(cqe->byte_cnt); 603 if ((cqe [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/ehea/ |
H A D | ehea_main.c | 519 static inline int ehea_check_cqe(struct ehea_cqe *cqe, int *rq_num) argument 521 *rq_num = (cqe->type & EHEA_CQE_TYPE_RQ) >> 5; 522 if ((cqe->status & EHEA_CQE_STAT_ERR_MASK) == 0) 524 if (((cqe->status & EHEA_CQE_STAT_ERR_TCP) != 0) && 525 (cqe->header_length == 0)) 531 struct sk_buff *skb, struct ehea_cqe *cqe) 533 int length = cqe->num_bytes_transfered - 4; /*remove CRC */ 540 if (cqe->status & EHEA_CQE_BLIND_CKSUM) { 542 skb->csum = csum_unfold(~cqe->inet_checksum_value); 549 struct ehea_cqe *cqe) 530 ehea_fill_skb(struct net_device *dev, struct sk_buff *skb, struct ehea_cqe *cqe) argument 547 get_skb_by_index(struct sk_buff **skb_array, int arr_len, struct ehea_cqe *cqe) argument 601 ehea_treat_poll_error(struct ehea_port_res *pr, int rq, struct ehea_cqe *cqe, int *processed_rq2, int *processed_rq3) argument 640 struct ehea_cqe *cqe = priv; local 668 ehea_proc_skb(struct ehea_port_res *pr, struct ehea_cqe *cqe, struct sk_buff *skb) argument 697 struct ehea_cqe *cqe; local 837 struct ehea_cqe *cqe; local 912 struct ehea_cqe *cqe; local [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/bnx2x/ |
H A D | bnx2x_cmn.c | 320 u16 queue, int pad, int len, union eth_rx_cqe *cqe, 339 (le16_to_cpu(cqe->fast_path_cqe.pars_flags.flags) & 379 &cqe->fast_path_cqe, cqe_idx)) { 384 le16_to_cpu(cqe->fast_path_cqe. 412 static inline void bnx2x_set_skb_rxhash(struct bnx2x *bp, union eth_rx_cqe *cqe, argument 417 (cqe->fast_path_cqe.status_flags & 420 le32_to_cpu(cqe->fast_path_cqe.rss_hash_result); 459 union eth_rx_cqe *cqe; local 474 cqe = &fp->rx_comp_ring[comp_ring_cons]; 475 cqe_fp_flags = cqe 319 bnx2x_tpa_stop(struct bnx2x *bp, struct bnx2x_fastpath *fp, u16 queue, int pad, int len, union eth_rx_cqe *cqe, u16 cqe_idx) argument [all...] |
H A D | bnx2x.h | 461 #define BNX2X_IP_CSUM_ERR(cqe) \ 462 (!((cqe)->fast_path_cqe.status_flags & \ 464 ((cqe)->fast_path_cqe.type_error_flags & \ 467 #define BNX2X_L4_CSUM_ERR(cqe) \ 468 (!((cqe)->fast_path_cqe.status_flags & \ 470 ((cqe)->fast_path_cqe.type_error_flags & \ 473 #define BNX2X_RX_CSUM_OK(cqe) \ 474 (!(BNX2X_L4_CSUM_ERR(cqe) || BNX2X_IP_CSUM_ERR(cqe))) 481 #define BNX2X_RX_SUM_FIX(cqe) \ [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/bnx2i/ |
H A D | bnx2i.h | 320 * @num_cqe_rcvd: statistic counter, total cqe's received 434 struct cqe { struct 578 struct cqe *cq_virt; 582 struct cqe *cq_prod_qe; 583 struct cqe *cq_cons_qe; 584 struct cqe *cq_first_qe; 585 struct cqe *cq_last_qe;
|
H A D | bnx2i_hwi.c | 1260 * @cqe: pointer to newly DMA'ed CQE entry for processing 1266 struct cqe *cqe) 1275 resp_cqe = (struct bnx2i_cmd_response *)cqe; 1300 resp_cqe = (struct bnx2i_cmd_response *)cqe; 1346 * @cqe: pointer to newly DMA'ed CQE entry for processing 1352 struct cqe *cqe) 1361 login = (struct bnx2i_login_response *) cqe; 1413 * @cqe 1264 bnx2i_process_scsi_cmd_resp(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1350 bnx2i_process_login_resp(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1417 bnx2i_process_tmf_resp(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1456 bnx2i_process_logout_resp(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1502 bnx2i_process_nopin_local_cmpl(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1543 bnx2i_process_nopin_mesg(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1596 bnx2i_process_async_mesg(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1646 bnx2i_process_reject_mesg(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument 1683 bnx2i_process_cmd_cleanup_resp(struct iscsi_session *session, struct bnx2i_conn *bnx2i_conn, struct cqe *cqe) argument [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/include/rdma/ |
H A D | ib_user_verbs.h | 255 __u32 cqe; member in struct:ib_uverbs_create_cq 264 __u32 cqe; member in struct:ib_uverbs_create_cq_resp 270 __u32 cqe; member in struct:ib_uverbs_resize_cq 275 __u32 cqe; member in struct:ib_uverbs_resize_cq_resp
|