/barrelfish-master/lib/devif/backends/net/mlx4/drivers/net/mlx4/ |
H A D | en_cq.c | 34 #include <linux/mlx4/cq.h> 43 static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event) { argument 50 struct mlx4_en_cq *cq; local 53 /*cq = kzalloc_node(sizeof(struct mlx4_en_cq), GFP_KERNEL, node); 54 if (!cq) {*/ 55 cq = calloc(1, sizeof(struct mlx4_en_cq)); 56 if (!cq) { 62 cq->size = entries; 63 cq->buf_size = cq 104 mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, int cq_idx) argument 200 mlx4_en_arm_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) argument [all...] |
H A D | cq.c | 41 #include <linux/mlx4/cq.h> 45 #include <linux/mlx4/cq.h> 68 struct mlx4_cq *cq; local 72 cq = radix_tree_lookup(&priv->cq_table.tree, 74 /*if (cq) 75 atomic_inc(&cq->refcount); 79 if (!cq) { 84 ++cq->arm_sn; 86 if (cq->comp) 87 cq 190 mlx4_cq_ignore_overrun(struct mlx4_dev *dev, struct mlx4_cq *cq) argument 279 mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, unsigned vector, int collapsed, int timestamp_en) argument [all...] |
H A D | en_tx.c | 33 #include <linux/mlx4/cq.h> 242 struct mlx4_en_tx_ring *ring, int cq, int user_prio) { 246 ring->cqn = cq; 393 struct mlx4_en_cq *cq = priv->tx_cq[0]; local 394 struct mlx4_cq *mcq = &cq->mcq; 395 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; 402 int size = cq->size; 404 struct mlx4_cqe *buf = cq->buf; 487 struct mlx4_en_cq *cq) { 489 struct mlx4_cq *mcq = &cq 241 mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int cq, int user_prio) argument 486 mlx4_en_process_tx_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) argument 609 struct mlx4_en_cq *cq = priv->tx_cq[tx_ind]; local 909 struct mlx4_en_cq *cq; local [all...] |
H A D | en_netdev.c | 48 #include <linux/mlx4/cq.h> 61 #include <linux/mlx4/cq.h> 84 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi); 85 struct net_device *dev = cq->dev; 87 struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring]; 93 if (!mlx4_en_cq_lock_poll(cq)) 96 done = mlx4_en_process_rx_cq(dev, cq, 4); 104 mlx4_en_cq_unlock_poll(cq); 1028 struct mlx4_en_cq *cq; 1033 cq 1589 struct mlx4_en_cq *cq; local [all...] |
H A D | resource_tracker.c | 192 struct res_cq *cq; 405 priv->dev.quotas.cq = priv->dev.caps.num_cqs - priv->dev.caps.reserved_cqs; 415 priv->dev.quotas.cq = priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[pf]; 1392 enum res_cq_states state, struct res_cq **cq) 1435 if (cq) 1436 *cq = r; 3079 struct res_cq *cq; 3082 err = cq_res_start_move_to(dev, slave, cqn, RES_CQ_HW, &cq); 3095 cq->mtt = mtt; 3115 struct res_cq *cq; [all...] |
H A D | en_rx.c | 35 #include <linux/mlx4/cq.h> 50 #include <linux/mlx4/cq.h> 569 struct mlx4_en_cq *cq = priv->rx_cq[0]; local 570 struct mlx4_cqe *buf = cq->buf; 572 struct mlx4_cq *mcq = &cq->mcq; 575 int size = cq->size; 583 // printf("cqe->owner_sr_opcode %d; cons_index %d:%d; size %d; cq->buf %p; index %d; qpn %x\n", 584 // cqe->owner_sr_opcode, cons_index, ring->prod, size, cq->buf, index, cqe->vlan_my_qpn); 637 int mlx4_en_process_rx_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, argument 641 struct mlx4_en_rx_ring *ring = priv->rx_ring[cq 798 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); local [all...] |
H A D | mlx4_en.h | 66 #include <linux/mlx4/cq.h> 379 /*Per-core Tx cq processing support*/ 636 static inline void mlx4_en_cq_init_lock(struct mlx4_en_cq *cq) 638 spin_lock_init(&cq->poll_lock); 639 cq->state = MLX4_EN_CQ_STATEIDLE; 642 called from the device poll rutine to get ownership of a cq 643 static inline bool mlx4_en_cq_lock_napi(struct mlx4_en_cq *cq) 646 spin_lock(&cq->poll_lock); 647 if (cq->state & MLX4_CQ_LOCKED) { 648 WARN_ON(cq [all...] |
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 168 static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry) argument 170 return get_cqe_from_buf(&cq->buf, entry); 178 static inline struct mthca_cqe *next_cqe_sw(struct mthca_cq *cq) argument 180 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); 200 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index 203 static inline void update_cons_index(struct mthca_dev *dev, struct mthca_cq *cq, argument 207 *cq->set_ci_db = cpu_to_be32(cq 223 struct mthca_cq *cq; local 240 struct mthca_cq *cq; local 277 mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn, struct mthca_srq *srq) argument 329 mthca_cq_resize_copy_cqes(struct mthca_cq *cq) argument 376 handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp *qp, int wqe_index, int is_send, struct mthca_err_cqe *cqe, struct ib_wc *entry, int *free_cqe) argument 482 mthca_poll_one(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp **cur_qp, int *freed, struct ib_wc *entry) argument 663 struct mthca_cq *cq = to_mcq(ibcq); local 728 mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) argument 743 struct mthca_cq *cq = to_mcq(ibcq); local 773 mthca_init_cq(struct mthca_dev *dev, int nent, struct mthca_ucontext *ctx, u32 pdn, struct mthca_cq *cq) argument 900 get_cq_refcount(struct mthca_dev *dev, struct mthca_cq *cq) argument 911 mthca_free_cq(struct mthca_dev *dev, struct mthca_cq *cq) argument [all...] |
H A D | mthca_provider.c | 681 struct mthca_cq *cq; local 705 cq = kmalloc(sizeof *cq, GFP_KERNEL); 706 if (!cq) { 712 cq->buf.mr.ibmr.lkey = ucmd.lkey; 713 cq->set_ci_db_index = ucmd.set_db_index; 714 cq->arm_db_index = ucmd.arm_db_index; 723 cq); 727 if (context && ib_copy_to_udata(udata, &cq->cqn, sizeof (__u32))) { 728 mthca_free_cq(to_mdev(ibdev), cq); local 752 mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq, int entries) argument 800 struct mthca_cq *cq = to_mcq(ibcq); local 876 mthca_destroy_cq(struct ib_cq *cq) argument [all...] |
H A D | mthca_dev.h | 242 struct mthca_array cq; member in struct:mthca_cq_table 496 int mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags); 497 int mthca_arbel_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags); 500 struct mthca_cq *cq); 502 struct mthca_cq *cq); 506 void mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn, 508 void mthca_cq_resize_copy_cqes(struct mthca_cq *cq);
|
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 35 #include <linux/mlx4/cq.h> 52 static void mlx4_ib_cq_comp(struct mlx4_cq *cq) { argument 53 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; 57 static void mlx4_ib_cq_event(struct mlx4_cq *cq, enum mlx4_event type) { argument 63 "on CQ %06x\n", type, cq->cqn); 67 ibcq = &to_mibcq(cq)->ibcq; 71 event.element.cq = ibcq; 80 static void *get_cqe(struct mlx4_ib_cq *cq, int n) { argument 81 return get_cqe_from_buf(&cq->buf, n); 84 static void *get_sw_cqe(struct mlx4_ib_cq *cq, in argument 92 next_cqe_sw(struct mlx4_ib_cq *cq) argument 178 struct mlx4_ib_cq *cq; local 460 struct mlx4_ib_cq *cq = to_mcq(ibcq); local 587 mlx4_ib_poll_one(struct mlx4_ib_cq *cq, struct mlx4_ib_qp **cur_qp, struct ib_wc *wc) argument 790 struct mlx4_ib_cq *cq = to_mcq(ibcq); local [all...] |
H A D | Makefile | 10 SRCS+= ah.c cq.c doorbell.c mad.c main.c mr.c qp.c srq.c wc.c
|
H A D | mlx4_ib.h | 107 struct ib_cq *cq; member in struct:mlx4_ib_xrcd 438 struct ib_cq *cq; member in struct:mlx4_ib_demux_pv_ctx 645 int mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period); 652 int mlx4_ib_destroy_cq(struct ib_cq *cq); 655 int mlx4_ib_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags); 657 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq); 658 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
|
/barrelfish-master/lib/libc/gen/ |
H A D | disklabel.c | 69 char *cp, *cq; /* can't be register */ local 81 cq = dp->d_typename; 83 while (cq < dp->d_typename + sizeof(dp->d_typename) - 1 && 84 (*cq = *cp) && *cq != '|' && *cq != ':') 85 cq++, cp++; 86 *cq = '\0'; 88 if (cgetstr(buf, "ty", &cq) > 0) { 89 if (strcmp(cq, "removabl [all...] |
/barrelfish-master/lib/devif/backends/net/mlx4/include/linux/mlx4/ |
H A D | cq.h | 147 static inline void mlx4_cq_arm(struct mlx4_cq *cq, u32 cmd, argument 153 sn = cq->arm_sn & 3; 154 ci = cq->cons_index & 0xffffff; 156 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci); 163 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn); 169 static inline void mlx4_cq_set_ci(struct mlx4_cq *cq) { argument 170 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); 177 int mlx4_cq_modify(struct mlx4_dev *dev, struct mlx4_cq *cq, 179 int mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq, [all...] |
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/ |
H A D | test_ib.c | 11 struct ib_cq *cq; local 33 cq = ib_create_cq(device, NULL, NULL, NULL, 1, 0); 34 if (!cq) 55 attr.send_cq = cq; 56 attr.recv_cq = cq; 155 poll_result = ib_poll_cq(cq, 1, &wc);
|
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/core/ |
H A D | verbs.c | 259 srq->ext.xrc.cq = NULL; 287 srq->ext.xrc.cq = xrc_cq; 328 xrc_cq = srq->ext.xrc.cq; 582 struct ib_cq *cq; local 584 cq = device->create_cq(device, cqe, comp_vector, NULL, NULL); 586 if (!IS_ERR(cq)) { 587 cq->device = device; 588 cq->uobject = NULL; 589 cq->comp_handler = comp_handler; 590 cq [all...] |
H A D | mad.c | 2189 ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP); 2191 while (ib_poll_cq(port_priv->cq, 1, &wc) == 1) { 2467 static void ib_mad_thread_completion_handler(struct ib_cq *cq, void *arg) { argument 2468 struct ib_mad_port_private *port_priv = cq->cq_context; 2635 ret = ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP); 2690 qp_init_attr.send_cq = qp_info->port_priv->cq; 2691 qp_init_attr.recv_cq = qp_info->port_priv->cq; 2755 port_priv->cq = ib_create_cq(port_priv->device, 2758 if (IS_ERR(port_priv->cq)) { 2760 ret = PTR_ERR(port_priv->cq); [all...] |
H A D | uverbs_cmd.c | 224 static void put_cq_read(struct ib_cq *cq) argument 226 put_uobj_read(cq->uobject); 767 struct ib_cq *cq; local 804 cq = file->device->ib_dev->create_cq(file->device->ib_dev, cmd.cqe, 807 if (IS_ERR(cq)) { 808 ret = PTR_ERR(cq); 812 cq->device = file->device->ib_dev; 813 cq->uobject = &obj->uobject; 814 cq->comp_handler = ib_uverbs_comp_handler; 815 cq 866 struct ib_cq *cq; local 902 struct ib_cq *cq; local 965 struct ib_cq *cq; local 989 struct ib_cq *cq; local [all...] |
H A D | mad_priv.h | 207 struct ib_cq *cq; member in struct:ib_mad_port_private
|
H A D | uverbs_main.c | 237 struct ib_cq *cq = uobj->object; local 238 struct ib_uverbs_event_file *ev_file = cq->cq_context; 243 ib_destroy_cq(cq); 411 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context) argument 433 uobj = container_of(cq->uobject, struct ib_ucq_object, uobject); 435 entry->desc.comp.cq_handle = cq->uobject->user_handle; 485 struct ib_ucq_object *uobj = container_of(event->element.cq->uobject,
|
/barrelfish-master/lib/libc/iconv/ |
H A D | citrus_mapper.c | 132 const char *cp, *cq; local 162 cq = _bcs_skip_nonws_len(cp, &len); 163 strlcpy(p, cp, (size_t)(cq - cp + 1)); 164 p += cq - cp + 1; 168 cp = _bcs_skip_ws_len(cq, &len);
|
/barrelfish-master/lib/devif/backends/net/mlx4/include/infiniband/ |
H A D | driver.h | 83 struct ibv_comp_channel *channel, int comp_vector, struct ibv_cq *cq, 86 int ibv_cmd_poll_cq(struct ibv_cq *cq, int ne, struct ibv_wc *wc); 87 int ibv_cmd_req_notify_cq(struct ibv_cq *cq, int solicited_only); 89 int ibv_cmd_resize_cq(struct ibv_cq *cq, int cqe, struct ibv_resize_cq *cmd, 91 int ibv_cmd_destroy_cq(struct ibv_cq *cq);
|
H A D | verbs.h | 221 struct ibv_cq *cq; member in union:ibv_async_event::__anon759 686 int (*poll_cq)(struct ibv_cq *cq, int num_entries, struct ibv_wc *wc); 687 int (*req_notify_cq)(struct ibv_cq *cq, int solicited_only); 688 void (*cq_event)(struct ibv_cq *cq); 689 int (*resize_cq)(struct ibv_cq *cq, int cqe); 690 int (*destroy_cq)(struct ibv_cq *cq); 872 * @cq: The CQ to resize. 875 * Users can examine the cq structure to determine the actual CQ size. 877 int ibv_resize_cq(struct ibv_cq *cq, int cqe); 882 int ibv_destroy_cq(struct ibv_cq *cq); [all...] |
/barrelfish-master/lib/devif/backends/net/mlx4/include/rdma/ |
H A D | ib_verbs.h | 395 struct ib_cq *cq; member in union:ib_event::__anon887 600 struct ib_cq *cq; 1008 typedef void (*ib_comp_handler)(struct ib_cq *cq, void *cq_context); 1032 struct ib_cq *cq; member in struct:ib_srq::__anon899::__anon900 1243 int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period); 1244 int (*destroy_cq)(struct ib_cq *cq); 1245 int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata); 1247 int (*poll_cq)(struct ib_cq *cq, int num_entries, struct ib_wc *wc); 1249 int (*peek_cq)(struct ib_cq *cq, int wc_cnt); 1251 int (*req_notify_cq)(struct ib_cq *cq, enu 1720 ib_poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) argument 1764 ib_req_notify_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) argument [all...] |