Searched refs:cq (Results 1 - 25 of 30) sorted by relevance

12

/barrelfish-master/lib/devif/backends/net/mlx4/drivers/net/mlx4/
H A Den_cq.c34 #include <linux/mlx4/cq.h>
43 static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event) { argument
50 struct mlx4_en_cq *cq; local
53 /*cq = kzalloc_node(sizeof(struct mlx4_en_cq), GFP_KERNEL, node);
54 if (!cq) {*/
55 cq = calloc(1, sizeof(struct mlx4_en_cq));
56 if (!cq) {
62 cq->size = entries;
63 cq->buf_size = cq
104 mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, int cq_idx) argument
200 mlx4_en_arm_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) argument
[all...]
H A Dcq.c41 #include <linux/mlx4/cq.h>
45 #include <linux/mlx4/cq.h>
68 struct mlx4_cq *cq; local
72 cq = radix_tree_lookup(&priv->cq_table.tree,
74 /*if (cq)
75 atomic_inc(&cq->refcount);
79 if (!cq) {
84 ++cq->arm_sn;
86 if (cq->comp)
87 cq
190 mlx4_cq_ignore_overrun(struct mlx4_dev *dev, struct mlx4_cq *cq) argument
279 mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, unsigned vector, int collapsed, int timestamp_en) argument
[all...]
H A Den_tx.c33 #include <linux/mlx4/cq.h>
242 struct mlx4_en_tx_ring *ring, int cq, int user_prio) {
246 ring->cqn = cq;
393 struct mlx4_en_cq *cq = priv->tx_cq[0]; local
394 struct mlx4_cq *mcq = &cq->mcq;
395 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring];
402 int size = cq->size;
404 struct mlx4_cqe *buf = cq->buf;
487 struct mlx4_en_cq *cq) {
489 struct mlx4_cq *mcq = &cq
241 mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, int cq, int user_prio) argument
486 mlx4_en_process_tx_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) argument
609 struct mlx4_en_cq *cq = priv->tx_cq[tx_ind]; local
909 struct mlx4_en_cq *cq; local
[all...]
H A Den_netdev.c48 #include <linux/mlx4/cq.h>
61 #include <linux/mlx4/cq.h>
84 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi);
85 struct net_device *dev = cq->dev;
87 struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring];
93 if (!mlx4_en_cq_lock_poll(cq))
96 done = mlx4_en_process_rx_cq(dev, cq, 4);
104 mlx4_en_cq_unlock_poll(cq);
1028 struct mlx4_en_cq *cq;
1033 cq
1589 struct mlx4_en_cq *cq; local
[all...]
H A Dresource_tracker.c192 struct res_cq *cq;
405 priv->dev.quotas.cq = priv->dev.caps.num_cqs - priv->dev.caps.reserved_cqs;
415 priv->dev.quotas.cq = priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[pf];
1392 enum res_cq_states state, struct res_cq **cq)
1435 if (cq)
1436 *cq = r;
3079 struct res_cq *cq;
3082 err = cq_res_start_move_to(dev, slave, cqn, RES_CQ_HW, &cq);
3095 cq->mtt = mtt;
3115 struct res_cq *cq;
[all...]
H A Den_rx.c35 #include <linux/mlx4/cq.h>
50 #include <linux/mlx4/cq.h>
569 struct mlx4_en_cq *cq = priv->rx_cq[0]; local
570 struct mlx4_cqe *buf = cq->buf;
572 struct mlx4_cq *mcq = &cq->mcq;
575 int size = cq->size;
583 // printf("cqe->owner_sr_opcode %d; cons_index %d:%d; size %d; cq->buf %p; index %d; qpn %x\n",
584 // cqe->owner_sr_opcode, cons_index, ring->prod, size, cq->buf, index, cqe->vlan_my_qpn);
637 int mlx4_en_process_rx_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, argument
641 struct mlx4_en_rx_ring *ring = priv->rx_ring[cq
798 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); local
[all...]
H A Dmlx4_en.h66 #include <linux/mlx4/cq.h>
379 /*Per-core Tx cq processing support*/
636 static inline void mlx4_en_cq_init_lock(struct mlx4_en_cq *cq)
638 spin_lock_init(&cq->poll_lock);
639 cq->state = MLX4_EN_CQ_STATEIDLE;
642 called from the device poll rutine to get ownership of a cq
643 static inline bool mlx4_en_cq_lock_napi(struct mlx4_en_cq *cq)
646 spin_lock(&cq->poll_lock);
647 if (cq->state & MLX4_CQ_LOCKED) {
648 WARN_ON(cq
[all...]
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c168 static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry) argument
170 return get_cqe_from_buf(&cq->buf, entry);
178 static inline struct mthca_cqe *next_cqe_sw(struct mthca_cq *cq) argument
180 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe));
200 * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index
203 static inline void update_cons_index(struct mthca_dev *dev, struct mthca_cq *cq, argument
207 *cq->set_ci_db = cpu_to_be32(cq
223 struct mthca_cq *cq; local
240 struct mthca_cq *cq; local
277 mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn, struct mthca_srq *srq) argument
329 mthca_cq_resize_copy_cqes(struct mthca_cq *cq) argument
376 handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp *qp, int wqe_index, int is_send, struct mthca_err_cqe *cqe, struct ib_wc *entry, int *free_cqe) argument
482 mthca_poll_one(struct mthca_dev *dev, struct mthca_cq *cq, struct mthca_qp **cur_qp, int *freed, struct ib_wc *entry) argument
663 struct mthca_cq *cq = to_mcq(ibcq); local
728 mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) argument
743 struct mthca_cq *cq = to_mcq(ibcq); local
773 mthca_init_cq(struct mthca_dev *dev, int nent, struct mthca_ucontext *ctx, u32 pdn, struct mthca_cq *cq) argument
900 get_cq_refcount(struct mthca_dev *dev, struct mthca_cq *cq) argument
911 mthca_free_cq(struct mthca_dev *dev, struct mthca_cq *cq) argument
[all...]
H A Dmthca_provider.c681 struct mthca_cq *cq; local
705 cq = kmalloc(sizeof *cq, GFP_KERNEL);
706 if (!cq) {
712 cq->buf.mr.ibmr.lkey = ucmd.lkey;
713 cq->set_ci_db_index = ucmd.set_db_index;
714 cq->arm_db_index = ucmd.arm_db_index;
723 cq);
727 if (context && ib_copy_to_udata(udata, &cq->cqn, sizeof (__u32))) {
728 mthca_free_cq(to_mdev(ibdev), cq); local
752 mthca_alloc_resize_buf(struct mthca_dev *dev, struct mthca_cq *cq, int entries) argument
800 struct mthca_cq *cq = to_mcq(ibcq); local
876 mthca_destroy_cq(struct ib_cq *cq) argument
[all...]
H A Dmthca_dev.h242 struct mthca_array cq; member in struct:mthca_cq_table
496 int mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags);
497 int mthca_arbel_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags);
500 struct mthca_cq *cq);
502 struct mthca_cq *cq);
506 void mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn,
508 void mthca_cq_resize_copy_cqes(struct mthca_cq *cq);
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mlx4/
H A Dcq.c35 #include <linux/mlx4/cq.h>
52 static void mlx4_ib_cq_comp(struct mlx4_cq *cq) { argument
53 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq;
57 static void mlx4_ib_cq_event(struct mlx4_cq *cq, enum mlx4_event type) { argument
63 "on CQ %06x\n", type, cq->cqn);
67 ibcq = &to_mibcq(cq)->ibcq;
71 event.element.cq = ibcq;
80 static void *get_cqe(struct mlx4_ib_cq *cq, int n) { argument
81 return get_cqe_from_buf(&cq->buf, n);
84 static void *get_sw_cqe(struct mlx4_ib_cq *cq, in argument
92 next_cqe_sw(struct mlx4_ib_cq *cq) argument
178 struct mlx4_ib_cq *cq; local
460 struct mlx4_ib_cq *cq = to_mcq(ibcq); local
587 mlx4_ib_poll_one(struct mlx4_ib_cq *cq, struct mlx4_ib_qp **cur_qp, struct ib_wc *wc) argument
790 struct mlx4_ib_cq *cq = to_mcq(ibcq); local
[all...]
H A DMakefile10 SRCS+= ah.c cq.c doorbell.c mad.c main.c mr.c qp.c srq.c wc.c
H A Dmlx4_ib.h107 struct ib_cq *cq; member in struct:mlx4_ib_xrcd
438 struct ib_cq *cq; member in struct:mlx4_ib_demux_pv_ctx
645 int mlx4_ib_modify_cq(struct ib_cq *cq, u16 cq_count, u16 cq_period);
652 int mlx4_ib_destroy_cq(struct ib_cq *cq);
655 int mlx4_ib_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags);
657 void __mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
658 void mlx4_ib_cq_clean(struct mlx4_ib_cq *cq, u32 qpn, struct mlx4_ib_srq *srq);
/barrelfish-master/lib/libc/gen/
H A Ddisklabel.c69 char *cp, *cq; /* can't be register */ local
81 cq = dp->d_typename;
83 while (cq < dp->d_typename + sizeof(dp->d_typename) - 1 &&
84 (*cq = *cp) && *cq != '|' && *cq != ':')
85 cq++, cp++;
86 *cq = '\0';
88 if (cgetstr(buf, "ty", &cq) > 0) {
89 if (strcmp(cq, "removabl
[all...]
/barrelfish-master/lib/devif/backends/net/mlx4/include/linux/mlx4/
H A Dcq.h147 static inline void mlx4_cq_arm(struct mlx4_cq *cq, u32 cmd, argument
153 sn = cq->arm_sn & 3;
154 ci = cq->cons_index & 0xffffff;
156 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci);
163 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn);
169 static inline void mlx4_cq_set_ci(struct mlx4_cq *cq) { argument
170 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff);
177 int mlx4_cq_modify(struct mlx4_dev *dev, struct mlx4_cq *cq,
179 int mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq,
[all...]
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/
H A Dtest_ib.c11 struct ib_cq *cq; local
33 cq = ib_create_cq(device, NULL, NULL, NULL, 1, 0);
34 if (!cq)
55 attr.send_cq = cq;
56 attr.recv_cq = cq;
155 poll_result = ib_poll_cq(cq, 1, &wc);
/barrelfish-master/lib/devif/backends/net/mlx4/drivers/infiniband/core/
H A Dverbs.c259 srq->ext.xrc.cq = NULL;
287 srq->ext.xrc.cq = xrc_cq;
328 xrc_cq = srq->ext.xrc.cq;
582 struct ib_cq *cq; local
584 cq = device->create_cq(device, cqe, comp_vector, NULL, NULL);
586 if (!IS_ERR(cq)) {
587 cq->device = device;
588 cq->uobject = NULL;
589 cq->comp_handler = comp_handler;
590 cq
[all...]
H A Dmad.c2189 ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP);
2191 while (ib_poll_cq(port_priv->cq, 1, &wc) == 1) {
2467 static void ib_mad_thread_completion_handler(struct ib_cq *cq, void *arg) { argument
2468 struct ib_mad_port_private *port_priv = cq->cq_context;
2635 ret = ib_req_notify_cq(port_priv->cq, IB_CQ_NEXT_COMP);
2690 qp_init_attr.send_cq = qp_info->port_priv->cq;
2691 qp_init_attr.recv_cq = qp_info->port_priv->cq;
2755 port_priv->cq = ib_create_cq(port_priv->device,
2758 if (IS_ERR(port_priv->cq)) {
2760 ret = PTR_ERR(port_priv->cq);
[all...]
H A Duverbs_cmd.c224 static void put_cq_read(struct ib_cq *cq) argument
226 put_uobj_read(cq->uobject);
767 struct ib_cq *cq; local
804 cq = file->device->ib_dev->create_cq(file->device->ib_dev, cmd.cqe,
807 if (IS_ERR(cq)) {
808 ret = PTR_ERR(cq);
812 cq->device = file->device->ib_dev;
813 cq->uobject = &obj->uobject;
814 cq->comp_handler = ib_uverbs_comp_handler;
815 cq
866 struct ib_cq *cq; local
902 struct ib_cq *cq; local
965 struct ib_cq *cq; local
989 struct ib_cq *cq; local
[all...]
H A Dmad_priv.h207 struct ib_cq *cq; member in struct:ib_mad_port_private
H A Duverbs_main.c237 struct ib_cq *cq = uobj->object; local
238 struct ib_uverbs_event_file *ev_file = cq->cq_context;
243 ib_destroy_cq(cq);
411 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context) argument
433 uobj = container_of(cq->uobject, struct ib_ucq_object, uobject);
435 entry->desc.comp.cq_handle = cq->uobject->user_handle;
485 struct ib_ucq_object *uobj = container_of(event->element.cq->uobject,
/barrelfish-master/lib/libc/iconv/
H A Dcitrus_mapper.c132 const char *cp, *cq; local
162 cq = _bcs_skip_nonws_len(cp, &len);
163 strlcpy(p, cp, (size_t)(cq - cp + 1));
164 p += cq - cp + 1;
168 cp = _bcs_skip_ws_len(cq, &len);
/barrelfish-master/lib/devif/backends/net/mlx4/include/infiniband/
H A Ddriver.h83 struct ibv_comp_channel *channel, int comp_vector, struct ibv_cq *cq,
86 int ibv_cmd_poll_cq(struct ibv_cq *cq, int ne, struct ibv_wc *wc);
87 int ibv_cmd_req_notify_cq(struct ibv_cq *cq, int solicited_only);
89 int ibv_cmd_resize_cq(struct ibv_cq *cq, int cqe, struct ibv_resize_cq *cmd,
91 int ibv_cmd_destroy_cq(struct ibv_cq *cq);
H A Dverbs.h221 struct ibv_cq *cq; member in union:ibv_async_event::__anon759
686 int (*poll_cq)(struct ibv_cq *cq, int num_entries, struct ibv_wc *wc);
687 int (*req_notify_cq)(struct ibv_cq *cq, int solicited_only);
688 void (*cq_event)(struct ibv_cq *cq);
689 int (*resize_cq)(struct ibv_cq *cq, int cqe);
690 int (*destroy_cq)(struct ibv_cq *cq);
872 * @cq: The CQ to resize.
875 * Users can examine the cq structure to determine the actual CQ size.
877 int ibv_resize_cq(struct ibv_cq *cq, int cqe);
882 int ibv_destroy_cq(struct ibv_cq *cq);
[all...]
/barrelfish-master/lib/devif/backends/net/mlx4/include/rdma/
H A Dib_verbs.h395 struct ib_cq *cq; member in union:ib_event::__anon887
600 struct ib_cq *cq;
1008 typedef void (*ib_comp_handler)(struct ib_cq *cq, void *cq_context);
1032 struct ib_cq *cq; member in struct:ib_srq::__anon899::__anon900
1243 int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period);
1244 int (*destroy_cq)(struct ib_cq *cq);
1245 int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata);
1247 int (*poll_cq)(struct ib_cq *cq, int num_entries, struct ib_wc *wc);
1249 int (*peek_cq)(struct ib_cq *cq, int wc_cnt);
1251 int (*req_notify_cq)(struct ib_cq *cq, enu
1720 ib_poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) argument
1764 ib_req_notify_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags) argument
[all...]

Completed in 270 milliseconds

12