Lines Matching refs:sg_list
73 if ((plen + wr->sg_list[i].length) < plen)
76 plen += wr->sg_list[i].length;
77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
110 if ((plen + wr->sg_list[i].length) < plen) {
113 plen += wr->sg_list[i].length;
115 cpu_to_be32(wr->sg_list[i].lkey);
117 cpu_to_be32(wr->sg_list[i].length);
119 cpu_to_be64(wr->sg_list[i].addr);
142 wqe->read.local_stag = cpu_to_be32(wr->sg_list[0].lkey);
143 wqe->read.local_len = cpu_to_be32(wr->sg_list[0].length);
144 wqe->read.local_to = cpu_to_be64(wr->sg_list[0].addr);
200 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list,
208 mhp = get_mhp(rhp, (sg_list[i].lkey) >> 8);
222 if (sg_list[i].addr < mhp->attr.va_fbo) {
226 if (sg_list[i].addr + ((u64) sg_list[i].length) <
227 sg_list[i].addr) {
231 if (sg_list[i].addr + ((u64) sg_list[i].length) >
236 offset = sg_list[i].addr - mhp->attr.va_fbo;
254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr,
264 wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
265 wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
268 wqe->recv.sgl[i].to = cpu_to_be64(((u32)wr->sg_list[i].addr) &
321 if (wr->sg_list[i].length > T3_STAG0_MAX_PBE_LEN)
328 if (wr->sg_list[i].lkey != 0)
331 wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
332 wqe->recv.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
497 if (wr->sg_list[0].lkey)