/linux-master/arch/powerpc/platforms/powernv/ |
H A D | opal-sensor-groups.c | 27 struct attribute_group sg; member in struct:sensor_group 129 static int __init add_attr_group(const __be32 *ops, int len, struct sensor_group *sg, argument 138 add_attr(handle, &sg->sgattrs[count], j); 139 sg->sg.attrs[count] = 140 &sg->sgattrs[count].attr.attr; 144 return sysfs_create_group(sg_kobj, &sg->sg); 162 struct device_node *sg, *node; local 165 sg [all...] |
/linux-master/lib/ |
H A D | scatterlist.c | 17 * @sg: The current sg entry 20 * Usually the next entry will be @sg@ + 1, but if this sg element is part 25 struct scatterlist *sg_next(struct scatterlist *sg) argument 27 if (sg_is_last(sg)) 30 sg++; 31 if (unlikely(sg_is_chain(sg))) 32 sg = sg_chain_ptr(sg); 47 sg_nents(struct scatterlist *sg) argument 70 sg_nents_for_len(struct scatterlist *sg, u64 len) argument 105 struct scatterlist *sg, *ret = NULL; local 140 sg_init_one(struct scatterlist *sg, const void *buf, unsigned int buflen) argument 171 sg_kfree(struct scatterlist *sg, unsigned int nents) argument 290 struct scatterlist *sg, *prv; local 615 struct scatterlist *sgl, *sg; local 687 struct scatterlist *sg; local 737 sg_page_count(struct scatterlist *sg) argument 761 sg_dma_page_count(struct scatterlist *sg) argument 814 struct scatterlist *sg; local 1113 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local 1168 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local 1214 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local 1274 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local [all...] |
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 254 struct gmap *sg, *next; local 260 list_for_each_entry_safe(sg, next, &gmap->children, list) { 261 list_del(&sg->list); 262 gmap_put(sg); 975 * Expected to be called with sg->mm->mmap_lock in read and 1021 * Expected to be called with sg->mm->mmap_lock in read 1057 * Called with sg->mm->mmap_lock in read. 1190 * @sg: pointer to the shadow guest address space structure 1194 * Called with the sg->guest_table_lock 1196 static inline void gmap_insert_rmap(struct gmap *sg, unsigne argument 1231 gmap_protect_rmap(struct gmap *sg, unsigned long raddr, unsigned long paddr, unsigned long len) argument 1312 gmap_unshadow_page(struct gmap *sg, unsigned long raddr) argument 1332 __gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr, unsigned long *pgt) argument 1349 gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr) argument 1379 __gmap_unshadow_sgt(struct gmap *sg, unsigned long raddr, unsigned long *sgt) argument 1407 gmap_unshadow_sgt(struct gmap *sg, unsigned long raddr) argument 1437 __gmap_unshadow_r3t(struct gmap *sg, unsigned long raddr, unsigned long *r3t) argument 1465 gmap_unshadow_r3t(struct gmap *sg, unsigned long raddr) argument 1495 __gmap_unshadow_r2t(struct gmap *sg, unsigned long raddr, unsigned long *r2t) argument 1523 gmap_unshadow_r2t(struct gmap *sg, unsigned long raddr) argument 1553 __gmap_unshadow_r1t(struct gmap *sg, unsigned long raddr, unsigned long *r1t) argument 1584 gmap_unshadow(struct gmap *sg) argument 1624 struct gmap *sg; local 1650 gmap_shadow_valid(struct gmap *sg, unsigned long asce, int edat_level) argument 1676 struct gmap *sg, *new; local 1766 gmap_shadow_r2t(struct gmap *sg, unsigned long saddr, unsigned long r2t, int fake) argument 1850 gmap_shadow_r3t(struct gmap *sg, unsigned long saddr, unsigned long r3t, int fake) argument 1934 gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt, int fake) argument 2018 gmap_shadow_pgt_lookup(struct gmap *sg, unsigned long saddr, unsigned long *pgt, int *dat_protection, int *fake) argument 2058 gmap_shadow_pgt(struct gmap *sg, unsigned long saddr, unsigned long pgt, int fake) argument 2137 gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte) argument 2205 gmap_shadow_notify(struct gmap *sg, unsigned long vmaddr, unsigned long gaddr) argument 2272 struct gmap *gmap, *sg, *next; local [all...] |
/linux-master/drivers/gpu/drm/loongson/ |
H A D | lsdc_gem.h | 15 struct sg_table *sg); 34 struct sg_table *sg,
|
/linux-master/drivers/crypto/gemini/ |
H A D | sl3516-ce-cipher.c | 32 struct scatterlist *sg; local 53 sg = areq->src; 54 while (sg) { 55 if ((sg->length % 16) != 0) { 59 if ((sg_dma_len(sg) % 16) != 0) { 63 if (!IS_ALIGNED(sg->offset, 16)) { 67 sg = sg_next(sg); 69 sg = areq->dst; 70 while (sg) { 135 struct scatterlist *sg; local [all...] |
/linux-master/net/core/ |
H A D | skmsg.c | 15 if (msg->sg.end > msg->sg.start && 16 elem_first_coalesce < msg->sg.end) 19 if (msg->sg.end < msg->sg.start && 20 (elem_first_coalesce > msg->sg.start || 21 elem_first_coalesce < msg->sg.end)) 31 u32 osize = msg->sg.size; 34 len -= msg->sg.size; 52 i = msg->sg [all...] |
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_wqe.h | 117 struct ib_sge *sg) 119 dseg->byte_count = cpu_to_be32(sg->length); 120 dseg->lkey = cpu_to_be32(sg->lkey); 121 dseg->addr = cpu_to_be64(sg->addr); 116 mthca_set_data_seg(struct mthca_data_seg *dseg, struct ib_sge *sg) argument
|
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_prime.h | 38 struct sg_table *sg);
|
/linux-master/include/linux/ |
H A D | page_reporting.h | 14 struct scatterlist *sg, unsigned int nents);
|
/linux-master/drivers/scsi/arm/ |
H A D | arm_scsi.h | 29 * copy_SCp_to_sg() Assumes contiguous allocation at @sg of at-most @max 31 * (possibly chained) sg-list 33 static inline int copy_SCp_to_sg(struct scatterlist *sg, struct scsi_pointer *SCp, int max) argument 42 sg_set_buf(sg, SCp->ptr, SCp->this_residual); 49 *(++sg) = *src_sg; 50 sg_mark_end(sg); 108 struct scatterlist *sg; local 111 scsi_for_each_sg(SCpnt, sg, sg_count, i) 112 len += sg->length;
|
/linux-master/drivers/dma-buf/ |
H A D | udmabuf.c | 29 struct sg_table *sg; member in struct:udmabuf 90 struct sg_table *sg; local 93 sg = kzalloc(sizeof(*sg), GFP_KERNEL); 94 if (!sg) 96 ret = sg_alloc_table_from_pages(sg, ubuf->pages, ubuf->pagecount, 101 ret = dma_map_sgtable(dev, sg, direction, 0); 104 return sg; 107 sg_free_table(sg); 108 kfree(sg); 112 put_sg_table(struct device *dev, struct sg_table *sg, enum dma_data_direction direction) argument 126 unmap_udmabuf(struct dma_buf_attachment *at, struct sg_table *sg, enum dma_data_direction direction) argument [all...] |
/linux-master/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem_dmabuf.c | 25 struct sg_table *sg; local 26 sg = omap_gem_get_sg(obj, dir); 27 if (IS_ERR(sg)) 28 return sg; 30 return sg; 34 struct sg_table *sg, enum dma_data_direction dir) 37 omap_gem_put_sg(obj, sg); 33 omap_gem_unmap_dma_buf(struct dma_buf_attachment *attachment, struct sg_table *sg, enum dma_data_direction dir) argument
|
/linux-master/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-cipher.c | 30 struct scatterlist *sg; local 45 sg = areq->src; 46 while (sg) { 47 todo = min(len, sg->length); 52 if (!IS_ALIGNED(sg->offset, 16)) { 57 sg = sg_next(sg); 60 sg = areq->dst; 61 while (sg) { 62 todo = min(len, sg 126 struct scatterlist *sg = areq->src; local 192 struct scatterlist *sg; local [all...] |
/linux-master/drivers/dma/ |
H A D | dma-axi-dmac.c | 139 struct axi_dmac_sg sg[] __counted_by(num_sgs); 232 struct axi_dmac_sg *sg; local 251 sg = &desc->sg[desc->num_submitted]; 254 if (sg->hw->id != AXI_DMAC_SG_UNUSED) { 255 sg->schedule_when_free = true; 272 sg->hw->id = axi_dmac_read(dmac, AXI_DMAC_REG_TRANSFER_ID); 276 axi_dmac_write(dmac, AXI_DMAC_REG_DEST_ADDRESS, sg->hw->dest_addr); 277 axi_dmac_write(dmac, AXI_DMAC_REG_DEST_STRIDE, sg->hw->dst_stride); 281 axi_dmac_write(dmac, AXI_DMAC_REG_SRC_ADDRESS, sg 318 axi_dmac_total_sg_bytes(struct axi_dmac_chan *chan, struct axi_dmac_sg *sg) argument 331 struct axi_dmac_sg *sg; local 378 struct axi_dmac_sg *sg; local 402 struct axi_dmac_sg *sg; local 582 axi_dmac_fill_linear_sg(struct axi_dmac_chan *chan, enum dma_transfer_direction direction, dma_addr_t addr, unsigned int num_periods, unsigned int period_len, struct axi_dmac_sg *sg) argument 631 struct scatterlist *sg; local [all...] |
/linux-master/drivers/target/tcm_fc/ |
H A D | tfc_io.c | 47 struct scatterlist *sg = NULL; local 81 sg = se_cmd->t_data_sg; 82 mem_len = sg->length; 83 mem_off = sg->offset; 84 page = sg_page(sg); 99 sg = sg_next(sg); 100 mem_len = min((size_t)sg->length, remaining); 101 mem_off = sg->offset; 102 page = sg_page(sg); 202 struct scatterlist *sg = NULL; local [all...] |
/linux-master/drivers/crypto/bcm/ |
H A D | util.h | 55 void __dump_sg(struct scatterlist *sg, unsigned int skip, unsigned int len); 57 #define dump_sg(sg, skip, len) __dump_sg(sg, skip, len) 77 static inline void dump_sg(struct scatterlist *sg, unsigned int skip, argument 84 int spu_sg_at_offset(struct scatterlist *sg, unsigned int skip, 87 /* Copy sg data, from skip, length len, to dest */
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 104 static void otx2_dma_unmap_skb_frags(struct otx2_nic *pfvf, struct sg_list *sg) argument 108 for (seg = 0; seg < sg->num_segs; seg++) { 109 otx2_dma_unmap_page(pfvf, sg->dma_addr[seg], 110 sg->size[seg], DMA_TO_DEVICE); 112 sg->num_segs = 0; 120 struct sg_list *sg; local 124 sg = &sq->sg[snd_comp->sqe_id]; 126 pa = otx2_iova_to_phys(pfvf->iommu_domain, sg->dma_addr[0]); 127 otx2_dma_unmap_page(pfvf, sg 143 struct sg_list *sg; local 257 struct nix_rx_sg_s *sg = &cqe->sg; local 344 struct nix_rx_sg_s *sg = &cqe->sg; local 617 struct nix_sqe_sg_s *sg = NULL; local 803 struct sg_list *sg = &sq->sg[sqe]; local 835 struct sg_list *sg = &sq->sg[sqe]; local 852 struct nix_sqe_sg_s *sg = NULL; local 1256 struct sg_list *sg; local 1321 struct sg_list *sg; local 1350 struct nix_sqe_sg_s *sg = NULL; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 120 struct sg_table *sg = NULL; local 153 sg = kmalloc(sizeof(*sg), GFP_KERNEL); 154 if (!sg) { 158 ret = sg_alloc_table(sg, 1, GFP_KERNEL); 162 sg_dma_address(sg->sgl) = dma_addr; 163 sg->sgl->length = adev->gart.table_size; 165 sg->sgl->dma_length = adev->gart.table_size; 180 bo->tbo.sg = sg; 229 struct sg_table *sg = adev->gart.bo->tbo.sg; local [all...] |
/linux-master/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_req.h | 474 * @sg: crypto request buffer. 482 struct scatterlist *sg; member in struct:nitrox_sgtable 567 * @sg: Destination SG list 573 static inline struct scatterlist *create_single_sg(struct scatterlist *sg, argument 576 sg_set_buf(sg, buf, buflen); 577 sg++; 578 return sg; 582 * create_multi_sg - Create multiple sg entries with buflen data length from 594 struct scatterlist *sg = to_sg; local 602 sg_set_buf(sg, sg_vir 647 struct scatterlist *sg; local 705 struct scatterlist *sg; local [all...] |
/linux-master/kernel/dma/ |
H A D | debug.h | 20 extern void debug_dma_map_sg(struct device *dev, struct scatterlist *sg, 51 struct scatterlist *sg, 55 struct scatterlist *sg, 70 static inline void debug_dma_map_sg(struct device *dev, struct scatterlist *sg, argument 119 struct scatterlist *sg, 125 struct scatterlist *sg, 118 debug_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, int direction) argument 124 debug_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, int direction) argument
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_bl.c | 67 struct scatterlist *sg; local 96 for_each_sg(sgl, sg, n, i) { 99 if (!sg->length) 102 if (left >= sg->length) { 103 left -= sg->length; 106 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left, 107 sg->length - left, 109 bufl->buffers[y].len = sg->length; 151 for_each_sg(sglout, sg, n_sglout, i) { 154 if (!sg 273 struct scatterlist *sg; local 364 qat_bl_realloc_map_new_dst(struct adf_accel_dev *accel_dev, struct scatterlist **sg, unsigned int dlen, struct qat_request_buffs *qat_bufs, gfp_t gfp) argument [all...] |
/linux-master/kernel/sched/ |
H A D | topology.c | 608 static void free_sched_groups(struct sched_group *sg, int free_sgc) argument 612 if (!sg) 615 first = sg; 617 tmp = sg->next; 619 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref)) 620 kfree(sg->sgc); 622 if (atomic_dec_and_test(&sg->ref)) 623 kfree(sg); 624 sg = tmp; 625 } while (sg ! 761 struct sched_group *sg = sd->groups; local 808 group_balance_cpu(struct sched_group *sg) argument 920 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) argument 959 struct sched_group *sg; local 980 init_overlap_sched_group(struct sched_domain *sd, struct sched_group *sg) argument 1036 struct sched_group *first = NULL, *last = NULL, *sg; local 1203 struct sched_group *sg; local 1259 struct sched_group *sg; local 1292 struct sched_group *sg = sd->groups; local 2235 struct sched_group *sg; local [all...] |
/linux-master/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_psfp.c | 132 const struct sparx5_psfp_sg *sg) 138 ips = sparx5_psfp_ipv_to_ips(sg->ipv); 139 base_lsb = sg->basetime.tv_sec & 0xffffffff; 140 base_msb = sg->basetime.tv_sec >> 32; 147 spx5_wr(sg->basetime.tv_nsec, sparx5, ANA_AC_SG_CONFIG_REG_1); 152 ANA_AC_SG_CONFIG_REG_3_LIST_LENGTH_SET(sg->num_entries) | 153 ANA_AC_SG_CONFIG_REG_3_INIT_GATE_STATE_SET(sg->gate_state) | 162 spx5_wr(sg->cycletime, sparx5, ANA_AC_SG_CONFIG_REG_4); 163 spx5_wr(sg->cycletimeext, sparx5, ANA_AC_SG_CONFIG_REG_5); 166 for (i = 0; i < sg 131 sparx5_psfp_sg_set(struct sparx5 *sparx5, u32 id, const struct sparx5_psfp_sg *sg) argument 226 sparx5_psfp_sg_add(struct sparx5 *sparx5, u32 uidx, struct sparx5_psfp_sg *sg, u32 *id) argument 253 const struct sparx5_psfp_sg sg = { 0 }; local [all...] |
/linux-master/arch/mips/jazz/ |
H A D | jazzdma.c | 547 struct scatterlist *sg; local 549 for_each_sg(sglist, sg, nents, i) { 551 arch_sync_dma_for_device(sg_phys(sg), sg->length, 553 sg->dma_address = vdma_alloc(sg_phys(sg), sg->length); 554 if (sg->dma_address == DMA_MAPPING_ERROR) 556 sg_dma_len(sg) = sg 566 struct scatterlist *sg; local 590 struct scatterlist *sg; local 600 struct scatterlist *sg; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_pages.c | 58 obj->mm.page_sizes.sg = 0; 61 obj->mm.page_sizes.sg |= BIT(i); 63 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg)); 231 obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0; 524 struct scatterlist *sg; local 531 /* As we iterate forward through the sg, we record each entry in a 545 /* We prefer to reuse the last sg so that repeated lookup of this 546 * (or the subsequent) sg are fast - comparing against the last 547 * sg is faster than going through the radixtree. 550 sg 634 struct scatterlist *sg; local 660 struct scatterlist *sg; local [all...] |