/freebsd-11.0-release/lib/libthr/ |
H A D | plockstat.d | 41 probe rw__acquire(void *lock, int wr); 42 probe rw__release(void *lock, int wr); 43 probe rw__block(void *lock, int wr); 44 probe rw__blocked(void *lock, int wr, int success); 45 probe rw__error(void *lock, int wr, int err);
|
/freebsd-11.0-release/contrib/flex/ |
H A D | tables.c | 56 int yytbl_write32 (struct yytbl_writer *wr, flex_uint32_t v); 57 int yytbl_write16 (struct yytbl_writer *wr, flex_uint16_t v); 58 int yytbl_write8 (struct yytbl_writer *wr, flex_uint8_t v); 59 int yytbl_writen (struct yytbl_writer *wr, void *v, flex_int32_t len); 68 * @param wr an uninitialized writer 72 int yytbl_writer_init (struct yytbl_writer *wr, FILE * out) 74 wr->out = out; 75 wr->total_written = 0; 127 static int yytbl_write_pad64 (struct yytbl_writer *wr) 131 pad = yypad64 (wr 71 yytbl_writer_init(struct yytbl_writer *wr, FILE * out) argument 126 yytbl_write_pad64(struct yytbl_writer *wr) argument 144 yytbl_hdr_fwrite(struct yytbl_writer *wr, const struct yytbl_hdr *th) argument 190 yytbl_data_fwrite(struct yytbl_writer *wr, struct yytbl_data *td) argument 268 yytbl_writen(struct yytbl_writer *wr, void *v, flex_int32_t len) argument 284 yytbl_write32(struct yytbl_writer *wr, flex_uint32_t v) argument 303 yytbl_write16(struct yytbl_writer *wr, flex_uint16_t v) argument 322 yytbl_write8(struct yytbl_writer *wr, flex_uint8_t v) argument [all...] |
H A D | tables.h | 72 int yytbl_hdr_fwrite (struct yytbl_writer *wr, 74 int yytbl_data_fwrite (struct yytbl_writer *wr, struct yytbl_data *td);
|
/freebsd-11.0-release/sys/dev/cxgbe/tom/ |
H A D | t4_tom_l2t.h | 44 t4_l2t_send(struct adapter *sc, struct wrqe *wr, struct l2t_entry *e) argument 47 t4_wrq_tx(sc, wr); 50 return (t4_l2t_send_slow(sc, wr, e));
|
H A D | t4_cpl_io.c | 77 struct wrqe *wr; local 91 wr = alloc_wrqe(roundup2(flowclen, 16), toep->ofld_txq); 92 if (wr == NULL) { 96 flowc = wrtod(wr); 97 memset(flowc, 0, wr->wr_len); 147 t4_wrq_tx(sc, wr); 153 struct wrqe *wr; local 177 wr = alloc_wrqe(sizeof(*req), toep->ofld_txq); 178 if (wr == NULL) { 182 req = wrtod(wr); 327 struct wrqe *wr; local 395 struct wrqe *wr; local 574 struct wrqe *wr; local 804 struct wrqe *wr; local 1219 struct wrqe *wr; local 1724 struct wrqe *wr; local [all...] |
/freebsd-11.0-release/contrib/ofed/libmthca/src/ |
H A D | qp.c | 101 int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, argument 126 for (nreq = 0; wr; ++nreq, wr = wr->next) { 129 *bad_wr = wr; 140 ((wr->send_flags & IBV_SEND_SIGNALED) ? 142 ((wr->send_flags & IBV_SEND_SOLICITED) ? 145 if (wr->opcode == IBV_WR_SEND_WITH_IMM || 146 wr->opcode == IBV_WR_RDMA_WRITE_WITH_IMM) 147 ((struct mthca_next_seg *) wqe)->imm = wr 334 mthca_tavor_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument 442 mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, struct ibv_send_wr **bad_wr) argument 707 mthca_arbel_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument [all...] |
H A D | srq.c | 81 struct ibv_recv_wr *wr, 99 for (nreq = 0; wr; wr = wr->next) { 106 *bad_wr = wr; 118 if (wr->num_sge > srq->max_gs) { 120 *bad_wr = wr; 125 for (i = 0; i < wr->num_sge; ++i) { 127 htonl(wr->sg_list[i].length); 129 htonl(wr 80 mthca_tavor_post_srq_recv(struct ibv_srq *ibsrq, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument 182 mthca_arbel_post_srq_recv(struct ibv_srq *ibsrq, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument [all...] |
/freebsd-11.0-release/contrib/ofed/libcxgb4/src/ |
H A D | qp.c | 87 struct ibv_send_wr *wr, int max, u32 *plenp) 95 for (i = 0; i < wr->num_sge; i++) { 96 if ((plen + wr->sg_list[i].length) > max) 98 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; 99 plen += wr->sg_list[i].length; 100 len = wr->sg_list[i].length; 142 struct ibv_send_wr *wr, u8 *len16) 148 if (wr->num_sge > T4_MAX_SEND_SGE) 150 if (wr->send_flags & IBV_SEND_SOLICITED) 161 if (wr 86 build_immd(struct t4_sq *sq, struct fw_ri_immd *immdp, struct ibv_send_wr *wr, int max, u32 *plenp) argument 141 build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, struct ibv_send_wr *wr, u8 *len16) argument 190 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, struct ibv_send_wr *wr, u8 *len16) argument 231 build_rdma_read(union t4_wr *wqe, struct ibv_send_wr *wr, u8 *len16) argument 258 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, struct ibv_recv_wr *wr, u8 *len16) argument 304 c4iw_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, struct ibv_send_wr **bad_wr) argument 406 c4iw_post_receive(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument [all...] |
/freebsd-11.0-release/contrib/ofed/libmlx4/src/ |
H A D | qp.c | 128 static void set_atomic_seg(struct mlx4_wqe_atomic_seg *aseg, struct ibv_send_wr *wr) argument 130 if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) { 131 aseg->swap_add = htonll(wr->wr.atomic.swap); 132 aseg->compare = htonll(wr->wr.atomic.compare_add); 134 aseg->swap_add = htonll(wr->wr.atomic.compare_add); 141 struct ibv_send_wr *wr) 143 memcpy(dseg->av, &to_mah(wr 140 set_datagram_seg(struct mlx4_wqe_datagram_seg *dseg, struct ibv_send_wr *wr) argument 189 mlx4_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, struct ibv_send_wr **bad_wr) argument 445 mlx4_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument [all...] |
H A D | srq.c | 65 struct ibv_recv_wr *wr, 77 for (nreq = 0; wr; ++nreq, wr = wr->next) { 78 if (wr->num_sge > srq->max_gs) { 80 *bad_wr = wr; 87 *bad_wr = wr; 91 srq->wrid[srq->head] = wr->wr_id; 97 for (i = 0; i < wr->num_sge; ++i) { 98 scat[i].byte_count = htonl(wr 64 mlx4_post_srq_recv(struct ibv_srq *ibsrq, struct ibv_recv_wr *wr, struct ibv_recv_wr **bad_wr) argument [all...] |
/freebsd-11.0-release/crypto/openssl/crypto/ripemd/asm/ |
H A D | rmd-586.pl | 39 @wr=( 5,14, 7, 0, 9, 2,11, 4,13, 6,15, 8, 1,10, 3,12, 446 &RIP5($A,$B,$C,$D,$E,$wr[ 0],$sr[ 0],$KR0,-2); 447 &RIP5($E,$A,$B,$C,$D,$wr[ 1],$sr[ 1],$KR0,0); 448 &RIP5($D,$E,$A,$B,$C,$wr[ 2],$sr[ 2],$KR0,0); 449 &RIP5($C,$D,$E,$A,$B,$wr[ 3],$sr[ 3],$KR0,0); 450 &RIP5($B,$C,$D,$E,$A,$wr[ 4],$sr[ 4],$KR0,0); 451 &RIP5($A,$B,$C,$D,$E,$wr[ 5],$sr[ 5],$KR0,0); 452 &RIP5($E,$A,$B,$C,$D,$wr[ 6],$sr[ 6],$KR0,0); 453 &RIP5($D,$E,$A,$B,$C,$wr[ 7],$sr[ 7],$KR0,0); 454 &RIP5($C,$D,$E,$A,$B,$wr[ [all...] |
/freebsd-11.0-release/sys/geom/ |
H A D | geom_vfs.h | 39 int g_vfs_open(struct vnode *vp, struct g_consumer **cpp, const char *fsname, int wr);
|
/freebsd-11.0-release/sys/dev/cxgb/ulp/iw_cxgb/ |
H A D | iw_cxgb_qp.c | 92 static int build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, argument 98 switch (wr->opcode) { 100 if (wr->send_flags & IB_SEND_SOLICITED) 107 if (wr->send_flags & IB_SEND_SOLICITED) 116 if (wr->num_sge > T3_MAX_SGE) 122 for (i = 0; i < wr->num_sge; i++) { 123 if ((plen + wr->sg_list[i].length) < plen) { 126 plen += wr->sg_list[i].length; 128 htobe32(wr->sg_list[i].lkey); 130 htobe32(wr 139 build_rdma_write(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 181 build_rdma_read(union t3_wr *wqe, struct ib_send_wr *wr, u8 *flit_cnt) argument 246 build_rdma_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 290 build_zero_stag_recv(struct iwch_qp *qhp, union t3_wr *wqe, struct ib_recv_wr *wr) argument 353 iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 452 iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument [all...] |
/freebsd-11.0-release/sys/ofed/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 1492 int ind, struct ib_send_wr *wr, 1502 mthca_ah_grh_present(to_mah(wr->wr.ud.ah)), 1506 err = mthca_read_ah(dev, to_mah(wr->wr.ud.ah), &sqp->ud_header); 1517 switch (wr->opcode) { 1525 sqp->ud_header.immediate_data = wr->ex.imm_data; 1534 sqp->ud_header.bth.solicited_event = !!(wr->send_flags & IB_SEND_SOLICITED); 1540 wr->wr 1491 build_mlx_header(struct mthca_dev *dev, struct mthca_sqp *sqp, int ind, struct ib_send_wr *wr, struct mthca_mlx_seg *mlx, struct mthca_data_seg *data) argument 1586 set_atomic_seg(struct mthca_atomic_seg *aseg, struct ib_send_wr *wr) argument 1599 set_tavor_ud_seg(struct mthca_tavor_ud_seg *useg, struct ib_send_wr *wr) argument 1609 set_arbel_ud_seg(struct mthca_arbel_ud_seg *useg, struct ib_send_wr *wr) argument 1617 mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 1820 mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument 1931 mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 2171 mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument [all...] |
H A D | mthca_srq.c | 494 int mthca_tavor_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, argument 513 for (nreq = 0; wr; wr = wr->next) { 521 *bad_wr = wr; 533 if (unlikely(wr->num_sge > srq->max_gs)) { 535 *bad_wr = wr; 540 for (i = 0; i < wr->num_sge; ++i) { 541 mthca_set_data_seg(wqe, wr->sg_list + i); 551 srq->wrid[ind] = wr 594 mthca_arbel_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument [all...] |
/freebsd-11.0-release/sys/ofed/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 2403 struct ib_send_wr *wr, 2410 struct mlx4_ib_ah *ah = to_mah(wr->wr.ud.ah); 2418 if (wr->opcode != IB_WR_SEND) 2423 for (i = 0; i < wr->num_sge; ++i) 2424 send_size += wr->sg_list[i].length; 2449 sqp->ud_header.bth.solicited_event = !!(wr->send_flags & IB_SEND_SOLICITED); 2453 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); 2510 static int build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, argument 2402 build_sriov_qp0_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, void *wqe, unsigned *mlx_seg_len) argument 2753 set_fmr_seg(struct mlx4_wqe_fmr_seg *fseg, struct ib_send_wr *wr) argument 2774 set_bind_seg(struct mlx4_wqe_bind_seg *bseg, struct ib_send_wr *wr) argument 2810 set_atomic_seg(struct mlx4_wqe_atomic_seg *aseg, struct ib_send_wr *wr) argument 2825 set_masked_atomic_seg(struct mlx4_wqe_masked_atomic_seg *aseg, struct ib_send_wr *wr) argument 2834 set_datagram_seg(struct mlx4_wqe_datagram_seg *dseg, struct ib_send_wr *wr) argument 2844 set_tunnel_datagram_seg(struct mlx4_ib_dev *dev, struct mlx4_wqe_datagram_seg *dseg, struct ib_send_wr *wr, enum ib_qp_type qpt) argument 2865 build_tunnel_header(struct ib_send_wr *wr, void *wqe, unsigned *mlx_seg_len) argument 2948 build_lso_seg(struct mlx4_wqe_lso_seg *wqe, struct ib_send_wr *wr, struct mlx4_ib_qp *qp, unsigned *lso_seg_len, __be32 *lso_hdr_sz, __be32 *blh) argument 2969 send_ieth(struct ib_send_wr *wr) argument 2991 lay_inline_data(struct mlx4_ib_qp *qp, struct ib_send_wr *wr, void *wqe, int *sz) argument 3073 mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 3417 mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument [all...] |
H A D | srq.c | 308 int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, argument 321 for (nreq = 0; wr; ++nreq, wr = wr->next) { 322 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { 324 *bad_wr = wr; 330 *bad_wr = wr; 334 srq->wrid[srq->head] = wr->wr_id; 340 for (i = 0; i < wr->num_sge; ++i) { 341 scat[i].byte_count = cpu_to_be32(wr [all...] |
H A D | mr.c | 686 struct ib_send_wr wr; local 690 memset(&wr, 0, sizeof(wr)); 691 wr.opcode = IB_WR_BIND_MW; 692 wr.wr_id = mw_bind->wr_id; 693 wr.send_flags = mw_bind->send_flags; 694 wr.wr.bind_mw.mw = mw; 695 wr.wr [all...] |
/freebsd-11.0-release/sys/dev/cxgbe/iw_cxgbe/ |
H A D | qp.c | 137 struct wrqe *wr; local 202 wr = alloc_wrqe(wr_len, &sc->sge.mgmtq); 203 if (wr == NULL) 205 res_wr = wrtod(wr); 262 t4_wrq_tx(sc, wr); 291 struct ib_send_wr *wr, int max, u32 *plenp) 299 for (i = 0; i < wr->num_sge; i++) { 300 if ((plen + wr->sg_list[i].length) > max) 302 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; 303 plen += wr 290 build_immd(struct t4_sq *sq, struct fw_ri_immd *immdp, struct ib_send_wr *wr, int max, u32 *plenp) argument 360 build_rdma_send(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 425 build_rdma_write(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 468 build_rdma_read(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 497 build_rdma_recv(struct c4iw_qp *qhp, union t4_recv_wr *wqe, struct ib_recv_wr *wr, u8 *len16) argument 512 build_fastreg(struct t4_sq *sq, union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 560 build_inv_stag(union t4_wr *wqe, struct ib_send_wr *wr, u8 *len16) argument 582 c4iw_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr, struct ib_send_wr **bad_wr) argument 698 c4iw_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr, struct ib_recv_wr **bad_wr) argument 906 struct wrqe *wr; local 1011 struct wrqe *wr; local 1095 struct wrqe *wr; local [all...] |
/freebsd-11.0-release/sys/dev/cxgbe/ |
H A D | offload.h | 42 INIT_ULPTX_WRH(&((w)->wr), wrlen, atomic, tid) 45 (w)->wr.wr_hi = htonl(V_FW_WR_OP(FW_TP_WR) | \ 46 V_FW_WR_IMMDLEN(sizeof(*w) - sizeof(w->wr))); \ 47 (w)->wr.wr_mid = htonl(V_FW_WR_LEN16(DIV_ROUND_UP(sizeof(*w), 16)) | \ 49 (w)->wr.wr_lo = cpu_to_be64(0); \
|
/freebsd-11.0-release/sys/geom/eli/ |
H A D | g_eli.c | 161 struct g_eli_worker *wr; local 167 LIST_FOREACH(wr, &sc->sc_workers, w_next) { 168 if (wr->w_number == bp->bio_pflags) 171 KASSERT(wr != NULL, ("Invalid worker (%u).", bp->bio_pflags)); 173 bp->bio_cmd == BIO_READ ? "READ" : "WRITE", (uintmax_t)wr->w_sid, 175 wr->w_sid = crp->crp_sid; 363 g_eli_newsession(struct g_eli_worker *wr) argument 369 sc = wr->w_softc; 392 error = crypto_newsession(&wr->w_sid, &crie, 396 error = crypto_newsession(&wr 427 g_eli_freesession(struct g_eli_worker *wr) argument 478 struct g_eli_worker *wr; local 696 struct g_eli_worker *wr; local [all...] |
/freebsd-11.0-release/sys/dev/iser/ |
H A D | iser_memory.c | 218 struct ib_send_wr *bad_wr, *wr = NULL; local 236 wr = &inv_wr; 243 fastreg_wr.wr.fast_reg.iova_start = rsc->frpl->page_list[0] + offset; 244 fastreg_wr.wr.fast_reg.page_list = rsc->frpl; 245 fastreg_wr.wr.fast_reg.page_list_len = plen; 246 fastreg_wr.wr.fast_reg.page_shift = SHIFT_4K; 247 fastreg_wr.wr.fast_reg.length = size; 248 fastreg_wr.wr.fast_reg.rkey = rsc->mr->rkey; 249 fastreg_wr.wr.fast_reg.access_flags = (IB_ACCESS_LOCAL_WRITE | 253 if (!wr) [all...] |
/freebsd-11.0-release/sys/boot/sparc64/loader/ |
H A D | locore.S | 34 wr %g0, FPRS_FEF, %fprs
|
/freebsd-11.0-release/contrib/unbound/util/storage/ |
H A D | slabhash.h | 123 * @param wr: set to true if you desire a writelock on the entry. 129 hashvalue_t hash, void* key, int wr); 180 * @param wr: if true, writelock is obtained, otherwise readlock. 184 void slabhash_traverse(struct slabhash* table, int wr,
|
/freebsd-11.0-release/contrib/bzip2/ |
H A D | compress.c | 125 Int32 wr; local 159 wr = 0; 165 AssertD ( wr <= i, "generateMTFValues(1)" ); 178 mtfv[wr] = BZ_RUNB; wr++; 181 mtfv[wr] = BZ_RUNA; wr++; 206 mtfv[wr] = j+1; wr++; s->mtfFreq[j+1]++; 216 mtfv[wr] [all...] |