/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/sound/soc/imx/ |
H A D | imx-pcm-dma-mx2.c | 36 struct scatterlist *sg_list; member in struct:imx_pcm_runtime_data 90 ret = imx_dma_setup_sg(iprtd->dma, iprtd->sg_list, iprtd->sg_count, 164 kfree(iprtd->sg_list); 166 iprtd->sg_list = kcalloc(iprtd->periods + 1, 168 if (!iprtd->sg_list) 173 sg_init_table(iprtd->sg_list, iprtd->sg_count); 177 iprtd->sg_list[i].page_link = 0; 178 iprtd->sg_list[i].offset = 0; 179 iprtd->sg_list[i].dma_address = dma_addr; 180 iprtd->sg_list[ [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/ehca/ |
H A D | ehca_qes.h | 156 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member in struct:ehca_wqe::__anon11586::__anon11587 164 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member in struct:ehca_wqe::__anon11586::__anon11588 168 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES - member in struct:ehca_wqe::__anon11586::__anon11589 176 struct ehca_vsgentry sg_list[MAX_WQE_SG_ENTRIES]; member in struct:ehca_wqe::__anon11586::__anon11590
|
H A D | ehca_reqs.c | 86 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); 92 wqe_p->u.all_rcv.sg_list[cnt_ds].vaddr = 93 recv_wr->sg_list[cnt_ds].addr; 94 wqe_p->u.all_rcv.sg_list[cnt_ds].lkey = 95 recv_wr->sg_list[cnt_ds].lkey; 96 wqe_p->u.all_rcv.sg_list[cnt_ds].length = 97 recv_wr->sg_list[cnt_ds].length; 120 struct ib_sge *sge = send_wr->sg_list; 176 memset(wqe_p, 0, offsetof(struct ehca_wqe, u.ud_av.sg_list)); 248 wqe_p->u.ud_av.sg_list[id [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_qp.c | 73 if ((plen + wr->sg_list[i].length) < plen) 76 plen += wr->sg_list[i].length; 77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); 78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); 79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); 110 if ((plen + wr->sg_list[i].length) < plen) { 113 plen += wr->sg_list[i].length; 115 cpu_to_be32(wr->sg_list[i].lkey); 117 cpu_to_be32(wr->sg_list[i].length); 119 cpu_to_be64(wr->sg_list[ 200 iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, u32 num_sgle, u32 * pbl_addr, u8 * page_size) argument [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/ipath/ |
H A D | ipath_ruc.c | 122 if (wqe->sg_list[i].length == 0) 125 if (!ipath_lkey_ok(qp, j ? &ss->sg_list[j - 1] : &ss->sge, 126 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 128 *lengthp += wqe->sg_list[i].length; 201 qp->r_sge.sg_list = qp->r_sg_list; 319 sqp->s_sge.sge = wqe->sg_list[0]; 320 sqp->s_sge.sg_list = wqe->sg_list + 1; 361 qp->r_sge.sge = wqe->sg_list[0]; 362 qp->r_sge.sg_list [all...] |
H A D | ipath_srq.c | 83 wqe->sg_list[i] = wr->sg_list[i]; 288 p->sg_list[i] = wqe->sg_list[i];
|
H A D | ipath_keys.c | 226 ss->sg_list = NULL; 263 ss->sg_list = NULL;
|
H A D | ipath_ud.c | 133 rsge.sg_list = qp->r_ud_sg_list; 183 sge = swqe->sg_list; 314 qp->s_sge.sge = wqe->sg_list[0]; 315 qp->s_sge.sg_list = wqe->sg_list + 1;
|
H A D | ipath_uc.c | 98 qp->s_sge.sge = wqe->sg_list[0]; 99 qp->s_sge.sg_list = wqe->sg_list + 1; 452 qp->r_sge.sg_list = NULL;
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/qib/ |
H A D | qib_ruc.c | 92 ss->sg_list = qp->r_sg_list; 95 if (wqe->sg_list[i].length == 0) 98 if (!qib_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, 99 &wqe->sg_list[i], IB_ACCESS_LOCAL_WRITE)) 101 qp->r_len += wqe->sg_list[i].length; 111 struct qib_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge; 422 sqp->s_sge.sge = wqe->sg_list[0]; 423 sqp->s_sge.sg_list = wqe->sg_list + 1; 460 qp->r_sge.sg_list [all...] |
H A D | qib_ud.c | 165 ssge.sg_list = swqe->sg_list + 1; 166 ssge.sge = *swqe->sg_list; 183 *sge = *ssge.sg_list++; 200 qp->r_sge.sge = *qp->r_sge.sg_list++; 310 qp->s_sge.sge = wqe->sg_list[0]; 311 qp->s_sge.sg_list = wqe->sg_list + 1; 565 qp->r_sge.sge = *qp->r_sge.sg_list++;
|
H A D | qib_uc.c | 99 qp->s_sge.sge = wqe->sg_list[0]; 100 qp->s_sge.sg_list = wqe->sg_list + 1; 291 qp->r_sge.sge = *qp->r_sge.sg_list++; 412 *qp->s_rdma_read_sge.sg_list++; 438 qp->r_sge.sg_list = NULL; 493 *qp->s_rdma_read_sge.sg_list++; 521 qp->r_sge.sge = *qp->r_sge.sg_list++;
|
H A D | qib_srq.c | 83 wqe->sg_list[i] = wr->sg_list[i]; 284 p->sg_list[i] = wqe->sg_list[i];
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/include/linux/ |
H A D | agp_backend.h | 85 struct scatterlist *sg_list; member in struct:agp_memory
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/qla2xxx/ |
H A D | qla_bsg.c | 168 bsg_job->reply_payload.sg_list, 196 sg_copy_to_buffer(bsg_job->request_payload.sg_list, 324 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, 331 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, 383 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, 385 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, 411 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, 418 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, 510 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, 512 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/message/i2o/ |
H A D | i2o_config.c | 528 struct i2o_dma sg_list[SG_TABLESIZE]; local 595 memset(sg_list, 0, sizeof(sg_list[0]) * SG_TABLESIZE); 628 p = &(sg_list[sg_index]); 701 sg_list[j].virt, sg_size)) { 704 c->name, sg_list[j].virt, 730 i2o_dma_free(&c->pdev->dev, &sg_list[i]); 774 struct i2o_dma sg_list[SG_TABLESIZE]; local 832 memset(sg_list, 0, sizeof(sg_list[ [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/include/scsi/ |
H A D | libiscsi_tcp.h | 114 struct scatterlist *sg_list, unsigned int sg_count,
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/aacraid/ |
H A D | commctrl.c | 478 void *sg_list[32]; local 503 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ 555 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { 614 sg_list[i] = p; // save so we can clean up later 665 sg_list[i] = p; // save so we can clean up later 716 sg_list[i] = p; // save so we can clean up later 752 sg_list[i] = p; // save so we can clean up later 792 if(copy_to_user(sg_user[i], sg_list[i], byte_count)){ 811 kfree(sg_list[ [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/net/ehea/ |
H A D | ehea_qmr.h | 119 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member in struct:ehea_swqe::__anon12834::__anon12835 128 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member in struct:ehea_swqe::__anon12834::__anon12836 145 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member in struct:ehea_rwqe
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/lpfc/ |
H A D | lpfc_bsg.c | 198 pci_unmap_sg(phba->pcidev, job->request_payload.sg_list, 200 pci_unmap_sg(phba->pcidev, job->reply_payload.sg_list, 313 request_nseg = pci_map_sg(phba->pcidev, job->request_payload.sg_list, 315 for_each_sg(job->request_payload.sg_list, sgel, request_nseg, numbde) { 325 reply_nseg = pci_map_sg(phba->pcidev, job->reply_payload.sg_list, 327 for_each_sg(job->reply_payload.sg_list, sgel, reply_nseg, numbde) { 387 pci_unmap_sg(phba->pcidev, job->request_payload.sg_list, 389 pci_unmap_sg(phba->pcidev, job->reply_payload.sg_list, 462 pci_unmap_sg(phba->pcidev, job->request_payload.sg_list, 464 pci_unmap_sg(phba->pcidev, job->reply_payload.sg_list, [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/scsi/ |
H A D | hptiop.h | 178 struct hpt_iopsg sg_list[1]; member in struct:hpt_iop_request_block_command 195 struct hpt_iopsg sg_list[1]; member in struct:hpt_iop_request_scsi_command
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/cxgb4/ |
H A D | qp.c | 248 if ((plen + wr->sg_list[i].length) > max) 250 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; 251 plen += wr->sg_list[i].length; 252 rem = wr->sg_list[i].length; 275 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, 284 if ((plen + sg_list[i].length) < plen) 286 plen += sg_list[i].length; 287 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | 288 sg_list[i].length); 291 *flitp = cpu_to_be64(sg_list[ 274 build_isgl(__be64 *queue_start, __be64 *queue_end, struct fw_ri_isgl *isglp, struct ib_sge *sg_list, int num_sge, u32 *plenp) argument [all...] |
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/hw/mlx4/ |
H A D | srq.c | 333 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); 334 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); 335 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr);
|
/netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/drivers/infiniband/core/ |
H A D | mad.c | 891 mad_send_wr->sg_list[0].length = hdr_len; 892 mad_send_wr->sg_list[0].lkey = mad_agent->mr->lkey; 893 mad_send_wr->sg_list[1].length = sizeof(struct ib_mad) - hdr_len; 894 mad_send_wr->sg_list[1].lkey = mad_agent->mr->lkey; 897 mad_send_wr->send_wr.sg_list = mad_send_wr->sg_list; 1012 sge = mad_send_wr->sg_list; 2109 mad_send_wr->sg_list[0].length, DMA_TO_DEVICE); 2112 mad_send_wr->sg_list[1].length, DMA_TO_DEVICE); 2532 struct ib_sge sg_list; local [all...] |
H A D | mad_priv.h | 132 struct ib_sge sg_list[IB_MAD_SEND_REQ_MAX_SG]; member in struct:ib_mad_send_wr_private
|