/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/linux/mtd/ |
H A D | nand.h | 243 * @wq: wait queue to sleep on if a NAND operation is in progress 249 wait_queue_head_t wq; member in struct:nand_hw_control 360 * @wq: [INTERN] wait queue to sleep on if a NAND operation is in progress
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/ap/gpl/iproute2/misc/ |
H A D | ss.c | 376 int rq, wq; member in struct:tcpstat 1161 &s.state, &s.wq, &s.rq, 1181 printf("%-6d %-6d ", s.rq, s.wq); 1769 &s.state, &s.wq, &s.rq, 1781 printf("%-6d %-6d ", s.rq, s.wq); 1881 int wq; member in struct:unixstat 1951 printf("%-6d %-6d ", s->rq, s->wq); 1990 &u->peer, &u->rq, &u->wq, &flags, &u->type, 2007 u->wq = 0; 2128 int rq, wq, r local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/i2c/busses/ |
H A D | i2c-ibm_iic.c | 333 wake_up_interruptible(&dev->wq); 414 ret = wait_event_interruptible_timeout(dev->wq, 696 init_waitqueue_head(&dev->wq);
|
H A D | i2c-ibm_iic.h | 48 wait_queue_head_t wq; member in struct:ibm_iic_private
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/core/ |
H A D | cm.c | 78 struct workqueue_struct *wq; member in struct:ib_cm 693 queue_delayed_work(cm.wq, &cm_id_priv->timewait_info->work.work, 3097 queue_delayed_work(cm.wq, &work->work, 0); 3200 queue_delayed_work(cm.wq, &work->work, 0); 3475 cm.wq = create_workqueue("ib_cm"); 3476 if (!cm.wq) 3485 destroy_workqueue(cm.wq); 3498 destroy_workqueue(cm.wq);
|
H A D | mad.c | 528 flush_workqueue(port_priv->wq); 777 queue_work(mad_agent_priv->qp_info->port_priv->wq, 1969 port_priv->wq, 2005 queue_delayed_work(mad_agent_priv->qp_info->port_priv->wq, 2466 port_priv->wq, 2499 queue_work(port_priv->wq, &port_priv->work); 2803 port_priv->wq = create_singlethread_workqueue(name); 2804 if (!port_priv->wq) { 2827 destroy_workqueue(port_priv->wq); 2866 destroy_workqueue(port_priv->wq); [all...] |
H A D | mad_priv.h | 209 struct workqueue_struct *wq; member in struct:ib_mad_port_private
|
H A D | mad_rmpp.c | 97 flush_workqueue(agent->qp_info->port_priv->wq); 448 queue_delayed_work(rmpp_recv->agent->qp_info->port_priv->wq, 541 queue_delayed_work(agent->qp_info->port_priv->wq,
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/cxgb3/ |
H A D | cxio_hal.c | 269 struct t3_wq *wq, struct cxio_ucontext *uctx) 271 int depth = 1UL << wq->size_log2; 272 int rqsize = 1UL << wq->rq_size_log2; 274 wq->qpid = get_qpid(rdev_p, uctx); 275 if (!wq->qpid) 278 wq->rq = kzalloc(depth * sizeof(u64), GFP_KERNEL); 279 if (!wq->rq) 282 wq->rq_addr = cxio_hal_rqtpool_alloc(rdev_p, rqsize); 283 if (!wq->rq_addr) 286 wq 268 cxio_create_qp(struct cxio_rdev *rdev_p, u32 kernel_domain, struct t3_wq *wq, struct cxio_ucontext *uctx) argument 329 cxio_destroy_qp(struct cxio_rdev *rdev_p, struct t3_wq *wq, struct cxio_ucontext *uctx) argument 343 insert_recv_cqe(struct t3_wq *wq, struct t3_cq *cq) argument 361 cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count) argument 375 insert_sq_cqe(struct t3_wq *wq, struct t3_cq *cq, struct t3_swsq *sqp) argument 396 cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count) argument 431 cqe_completes_wr(struct t3_cqe *cqe, struct t3_wq *wq) argument 449 cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument 466 cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count) argument 1028 flush_completed_wrs(struct t3_wq *wq, struct t3_cq *cq) argument 1057 create_read_req_cqe(struct t3_wq *wq, struct t3_cqe *hw_cqe, struct t3_cqe *read_cqe) argument 1071 advance_oldest_read(struct t3_wq *wq) argument 1102 cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe, u8 *cqe_flushed, u64 *cookie, u32 *credit) argument [all...] |
H A D | cxio_hal.h | 151 int cxio_create_qp(struct cxio_rdev *rdev, u32 kernel_domain, struct t3_wq *wq, 153 int cxio_destroy_qp(struct cxio_rdev *rdev, struct t3_wq *wq, 175 void cxio_flush_rq(struct t3_wq *wq, struct t3_cq *cq, int count); 176 void cxio_flush_sq(struct t3_wq *wq, struct t3_cq *cq, int count); 177 void cxio_count_rcqes(struct t3_cq *cq, struct t3_wq *wq, int *count); 178 void cxio_count_scqes(struct t3_cq *cq, struct t3_wq *wq, int *count); 180 int cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe,
|
H A D | cxio_wr.h | 613 u32 size_log2; /* total wq size */ 645 static inline void cxio_set_wq_in_error(struct t3_wq *wq) argument 647 wq->queue->flit[13] = 1;
|
H A D | iwch_cq.c | 49 struct t3_wq *wq; local 62 wq = NULL; 65 wq = &(qhp->wq); 67 ret = cxio_poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, 181 if (wq)
|
H A D | iwch_ev.c | 65 qhp->attr.state, qhp->wq.qpid, CQE_STATUS(rsp_msg->cqe)); 135 __FUNCTION__, qhp->wq.qpid, qhp->ep); 139 qhp->wq.qpid); 220 CQE_STATUS(rsp_msg->cqe), qhp->wq.qpid);
|
H A D | iwch_provider.c | 746 remove_handle(rhp, &rhp->qpidr, qhp->wq.qpid); 753 cxio_destroy_qp(&rhp->rdev, &qhp->wq, 757 ib_qp, qhp->wq.qpid, qhp); 812 qhp->wq.size_log2 = ilog2(wqsize); 813 qhp->wq.rq_size_log2 = ilog2(rqsize); 814 qhp->wq.sq_size_log2 = ilog2(sqsize); 816 if (cxio_create_qp(&rhp->rdev, !udata, &qhp->wq, 844 insert_handle(rhp, &rhp->qpidr, qhp, qhp->wq.qpid); 863 uresp.qpid = qhp->wq.qpid; 864 uresp.size_log2 = qhp->wq [all...] |
H A D | iwch_provider.h | 162 struct t3_wq wq; member in struct:iwch_qp
|
H A D | iwch_qp.c | 254 num_wrs = Q_FREECNT(qhp->wq.sq_rptr, qhp->wq.sq_wptr, 255 qhp->wq.sq_size_log2); 266 idx = Q_PTR2IDX(qhp->wq.wptr, qhp->wq.size_log2); 267 wqe = (union t3_wr *) (qhp->wq.queue + idx); 275 sqp = qhp->wq.sq + 276 Q_PTR2IDX(qhp->wq.sq_wptr, qhp->wq.sq_size_log2); 295 if (!qhp->wq [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/ipath/ |
H A D | ipath_qp.c | 366 if (qp->r_rq.wq) { 367 qp->r_rq.wq->head = 0; 368 qp->r_rq.wq->tail = 0; 430 if (qp->r_rq.wq) { 431 struct ipath_rwq *wq; local 438 wq = qp->r_rq.wq; 439 head = wq->head; 442 tail = wq->tail; 452 wq 675 struct ipath_rwq *wq = qp->r_rq.wq; local [all...] |
H A D | ipath_ruc.c | 160 struct ipath_rwq *wq; local 178 wq = rq->wq; 179 tail = wq->tail; 184 if (unlikely(tail == wq->head)) { 194 wq->tail = tail; 205 n = wq->head;
|
H A D | ipath_srq.c | 51 struct ipath_rwq *wq; local 67 wq = srq->rq.wq; 68 next = wq->head + 1; 71 if (next == wq->tail) { 78 wqe = get_rwqe_ptr(&srq->rq, wq->head); 83 wq->head = next; 131 srq->rq.wq = vmalloc_user(sizeof(struct ipath_rwq) + srq->rq.size * sz); 132 if (!srq->rq.wq) { 148 srq->rq.wq); 216 struct ipath_rwq *wq; local [all...] |
H A D | ipath_ud.c | 110 struct ipath_rwq *wq; local 170 wq = rq->wq; 171 tail = wq->tail; 173 if (unlikely(tail == wq->head)) { 183 wq->tail = tail; 191 wq->tail = tail; 200 n = wq->head;
|
H A D | ipath_verbs.c | 283 struct ipath_rwq *wq = qp->r_rq.wq; local 288 if (!(ib_ipath_state_ops[qp->state] & IPATH_POST_RECV_OK) || !wq) { 306 next = wq->head + 1; 309 if (next == wq->tail) { 316 wqe = get_rwqe_ptr(&qp->r_rq, wq->head); 321 wq->head = next;
|
H A D | ipath_verbs.h | 282 * Note that the wq array elements are variable size so you can't 289 struct ipath_rwqe wq[0]; member in struct:ipath_rwq 293 struct ipath_rwq *wq; member in struct:ipath_rq 438 * struct ipath_rwq.wq. This function does the array index computation. 444 ((char *) rq->wq->wq +
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 312 struct mlx4_ib_wq *wq; local 355 wq = &(*cur_qp)->sq; 357 wq->tail += (u16) (wqe_ctr - (u16) wq->tail); 358 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; 359 ++wq->tail; 366 wq = &(*cur_qp)->rq; 367 wc->wr_id = wq [all...] |
H A D | qp.c | 1169 static int mlx4_wq_overflow(struct mlx4_ib_wq *wq, int nreq, struct ib_cq *ib_cq) argument 1174 cur = wq->head - wq->tail; 1175 if (likely(cur + nreq < wq->max_post)) 1180 cur = wq->head - wq->tail; 1183 return cur + nreq >= wq->max_post;
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 494 struct mthca_wq *wq; local 543 wq = &(*cur_qp)->sq; 545 >> wq->wqe_shift); 551 wq = NULL; 557 wq = &(*cur_qp)->rq; 559 wqe_index = wqe >> wq->wqe_shift; 566 wqe_index = wq->max - 1; 570 if (wq) { 571 if (wq->last_comp < wqe_index) 572 wq [all...] |