/freebsd-10-stable/contrib/ofed/libmthca/src/ |
H A D | srq.c | 48 return srq->buf.buf + (n << srq->wqe_shift); 73 last_free->nda_op = htonl((ind << srq->wqe_shift) | 1); 150 doorbell[0] = htonl(first_ind << srq->wqe_shift); 166 doorbell[0] = htonl(first_ind << srq->wqe_shift); 268 for (srq->wqe_shift = 6; 1 << srq->wqe_shift < size; ++srq->wqe_shift) 271 srq->buf_size = srq->max << srq->wqe_shift; 295 next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); 302 (void *) scatter < wqe + (1 << srq->wqe_shift); [all...] |
H A D | qp.c | 63 return qp->buf.buf + (n << qp->rq.wqe_shift); 68 return qp->buf.buf + qp->send_wqe_offset + (n << qp->sq.wqe_shift); 292 htonl(((ind << qp->sq.wqe_shift) + 320 doorbell[0] = htonl(((qp->sq.next_ind << qp->sq.wqe_shift) + 404 doorbell[0] = htonl((qp->rq.next_ind << qp->rq.wqe_shift) | size0); 423 doorbell[0] = htonl((qp->rq.next_ind << qp->rq.wqe_shift) | size0); 658 htonl(((ind << qp->sq.wqe_shift) + 802 for (qp->rq.wqe_shift = 6; 1 << qp->rq.wqe_shift < size; 803 qp->rq.wqe_shift [all...] |
H A D | cq.c | 341 wqe_index = ((ntohl(cqe->wqe) - (*cur_qp)->send_wqe_offset) >> wq->wqe_shift); 347 wqe_index = wqe >> srq->wqe_shift; 354 wqe_index = wqe >> wq->wqe_shift; 577 ntohl(cqe->wqe) >> srq->wqe_shift);
|
H A D | mthca.h | 179 int wqe_shift; member in struct:mthca_srq 199 int wqe_shift; member in struct:mthca_wq
|
/freebsd-10-stable/contrib/ofed/libmlx4/src/ |
H A D | qp.c | 60 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); 65 return qp->buf.buf + qp->sq.offset + (n << qp->sq.wqe_shift); 99 ctrl->fence_size = 1 << (qp->sq.wqe_shift - 4); 581 for (qp->sq.wqe_shift = 6; 1 << qp->sq.wqe_shift < size; 582 qp->sq.wqe_shift++) 603 for (qp->rq.wqe_shift = 4; 604 1 << qp->rq.wqe_shift < qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg); 605 qp->rq.wqe_shift++) 608 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) [all...] |
H A D | srq.c | 48 return srq->buf.buf + (n << srq->wqe_shift); 143 for (srq->wqe_shift = 5; 1 << srq->wqe_shift < size; ++srq->wqe_shift) 146 buf_size = srq->max << srq->wqe_shift; 166 (void *) scatter < (void *) next + (1 << srq->wqe_shift);
|
H A D | mlx4.h | 238 int wqe_shift; member in struct:mlx4_srq 253 int wqe_shift; member in struct:mlx4_wq
|
H A D | verbs.c | 470 qp->sq_spare_wqes = (2048 >> qp->sq.wqe_shift) + 1; 505 cmd.log_sq_stride = qp->sq.wqe_shift;
|
/freebsd-10-stable/sys/ofed/drivers/infiniband/hw/mthca/ |
H A D | mthca_srq.c | 75 return srq->queue.direct.buf + (n << srq->wqe_shift); 77 return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf + 78 ((n << srq->wqe_shift) & (PAGE_SIZE - 1)); 102 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); 131 context->logstride_usrpage = cpu_to_be32((srq->wqe_shift - 4) << 29); 142 mthca_buf_free(dev, srq->max << srq->wqe_shift, &srq->queue, 162 err = mthca_buf_alloc(dev, srq->max << srq->wqe_shift, 182 next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); 189 (void *) scatter < wqe + (1 << srq->wqe_shift); 228 srq->wqe_shift [all...] |
H A D | mthca_provider.h | 228 int wqe_shift; member in struct:mthca_srq 254 int wqe_shift; member in struct:mthca_wq
|
H A D | mthca_qp.c | 210 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); 212 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + 213 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); 220 (n << qp->sq.wqe_shift); 223 (n << qp->sq.wqe_shift)) >> 225 ((qp->send_wqe_offset + (n << qp->sq.wqe_shift)) & 607 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; 611 qp_context->sq_size_stride |= qp->sq.wqe_shift - 4; 960 1 << qp->sq.wqe_shift)); 967 (min(dev->limits.max_desc_sz, 1 << qp->rq.wqe_shift) [all...] |
H A D | mthca_cq.c | 540 >> wq->wqe_shift); 546 wqe_index = wqe >> srq->wqe_shift; 553 wqe_index = wqe >> wq->wqe_shift;
|
/freebsd-10-stable/sys/ofed/drivers/infiniband/hw/mlx4/ |
H A D | srq.c | 43 return mlx4_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); 104 srq->msrq.wqe_shift = ilog2(desc_size); 284 mlx4_buf_free(dev->dev, msrq->msrq.max << msrq->msrq.wqe_shift,
|
H A D | qp.c | 198 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); 203 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); 226 s = roundup(size, 1U << qp->sq.wqe_shift); 228 ind = (i >> qp->sq.wqe_shift) + n; 232 wqe = buf + (i & ((1 << qp->sq.wqe_shift) - 1)); 287 post_nop_wqe(qp, ind, s << qp->sq.wqe_shift); 405 qp->rq.wqe_shift = ilog2(qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg)); 453 * request can include multiple units of 1 << wqe_shift. This 475 * We look for the smallest value of wqe_shift such that the 487 qp->sq.wqe_shift [all...] |
H A D | mlx4_ib.h | 158 int wqe_shift; member in struct:mlx4_ib_wq
|
/freebsd-10-stable/sys/dev/mlx5/mlx5_ib/ |
H A D | mlx5_ib_srq.c | 43 return mlx5_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); 282 srq->msrq.wqe_shift = ilog2(desc_size); 301 flgs = ((srq->msrq.wqe_shift - 4) | (is_xrc << 5) | (srq->wq_sig << 7)) << 24;
|
H A D | mlx5_ib_qp.c | 116 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); 274 qp->rq.wqe_shift = 0; 280 qp->rq.wqe_shift = ucmd->rq_wqe_shift; 281 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; 297 qp->rq.wqe_shift = ilog2(wqe_size); 298 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; 419 qp->sq.wqe_shift = ilog2(MLX5_SEND_WQE_BB); 439 int desc_sz = 1 << qp->sq.wqe_shift; 464 qp->buf_size = qp->rq.wqe_cnt << qp->rq.wqe_shift; 467 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) [all...] |
H A D | mlx5_ib.h | 175 int wqe_shift; member in struct:mlx5_ib_wq
|
/freebsd-10-stable/sys/ofed/drivers/net/mlx4/ |
H A D | srq.c | 194 srq_context->logstride = srq->wqe_shift - 4;
|
/freebsd-10-stable/sys/dev/mlx5/ |
H A D | driver.h | 447 int wqe_shift; member in struct:mlx5_core_srq
|
/freebsd-10-stable/sys/ofed/include/linux/mlx4/ |
H A D | device.h | 692 int wqe_shift; member in struct:mlx4_srq
|