Searched refs:sg (Results 26 - 50 of 896) sorted by relevance

1234567891011>>

/linux-master/arch/powerpc/platforms/powernv/
H A Dopal-sensor-groups.c27 struct attribute_group sg; member in struct:sensor_group
129 static int __init add_attr_group(const __be32 *ops, int len, struct sensor_group *sg, argument
138 add_attr(handle, &sg->sgattrs[count], j);
139 sg->sg.attrs[count] =
140 &sg->sgattrs[count].attr.attr;
144 return sysfs_create_group(sg_kobj, &sg->sg);
162 struct device_node *sg, *node; local
165 sg
[all...]
/linux-master/lib/
H A Dscatterlist.c17 * @sg: The current sg entry
20 * Usually the next entry will be @sg@ + 1, but if this sg element is part
25 struct scatterlist *sg_next(struct scatterlist *sg) argument
27 if (sg_is_last(sg))
30 sg++;
31 if (unlikely(sg_is_chain(sg)))
32 sg = sg_chain_ptr(sg);
47 sg_nents(struct scatterlist *sg) argument
70 sg_nents_for_len(struct scatterlist *sg, u64 len) argument
105 struct scatterlist *sg, *ret = NULL; local
140 sg_init_one(struct scatterlist *sg, const void *buf, unsigned int buflen) argument
171 sg_kfree(struct scatterlist *sg, unsigned int nents) argument
290 struct scatterlist *sg, *prv; local
615 struct scatterlist *sgl, *sg; local
687 struct scatterlist *sg; local
737 sg_page_count(struct scatterlist *sg) argument
761 sg_dma_page_count(struct scatterlist *sg) argument
814 struct scatterlist *sg; local
1113 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local
1168 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local
1214 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local
1274 struct scatterlist *sg = sgtable->sgl + sgtable->nents; local
[all...]
/linux-master/arch/s390/mm/
H A Dgmap.c254 struct gmap *sg, *next; local
260 list_for_each_entry_safe(sg, next, &gmap->children, list) {
261 list_del(&sg->list);
262 gmap_put(sg);
975 * Expected to be called with sg->mm->mmap_lock in read and
1021 * Expected to be called with sg->mm->mmap_lock in read
1057 * Called with sg->mm->mmap_lock in read.
1190 * @sg: pointer to the shadow guest address space structure
1194 * Called with the sg->guest_table_lock
1196 static inline void gmap_insert_rmap(struct gmap *sg, unsigne argument
1231 gmap_protect_rmap(struct gmap *sg, unsigned long raddr, unsigned long paddr, unsigned long len) argument
1312 gmap_unshadow_page(struct gmap *sg, unsigned long raddr) argument
1332 __gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr, unsigned long *pgt) argument
1349 gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr) argument
1379 __gmap_unshadow_sgt(struct gmap *sg, unsigned long raddr, unsigned long *sgt) argument
1407 gmap_unshadow_sgt(struct gmap *sg, unsigned long raddr) argument
1437 __gmap_unshadow_r3t(struct gmap *sg, unsigned long raddr, unsigned long *r3t) argument
1465 gmap_unshadow_r3t(struct gmap *sg, unsigned long raddr) argument
1495 __gmap_unshadow_r2t(struct gmap *sg, unsigned long raddr, unsigned long *r2t) argument
1523 gmap_unshadow_r2t(struct gmap *sg, unsigned long raddr) argument
1553 __gmap_unshadow_r1t(struct gmap *sg, unsigned long raddr, unsigned long *r1t) argument
1584 gmap_unshadow(struct gmap *sg) argument
1624 struct gmap *sg; local
1650 gmap_shadow_valid(struct gmap *sg, unsigned long asce, int edat_level) argument
1676 struct gmap *sg, *new; local
1766 gmap_shadow_r2t(struct gmap *sg, unsigned long saddr, unsigned long r2t, int fake) argument
1850 gmap_shadow_r3t(struct gmap *sg, unsigned long saddr, unsigned long r3t, int fake) argument
1934 gmap_shadow_sgt(struct gmap *sg, unsigned long saddr, unsigned long sgt, int fake) argument
2018 gmap_shadow_pgt_lookup(struct gmap *sg, unsigned long saddr, unsigned long *pgt, int *dat_protection, int *fake) argument
2058 gmap_shadow_pgt(struct gmap *sg, unsigned long saddr, unsigned long pgt, int fake) argument
2137 gmap_shadow_page(struct gmap *sg, unsigned long saddr, pte_t pte) argument
2205 gmap_shadow_notify(struct gmap *sg, unsigned long vmaddr, unsigned long gaddr) argument
2272 struct gmap *gmap, *sg, *next; local
[all...]
/linux-master/drivers/gpu/drm/loongson/
H A Dlsdc_gem.h15 struct sg_table *sg);
34 struct sg_table *sg,
/linux-master/drivers/crypto/gemini/
H A Dsl3516-ce-cipher.c32 struct scatterlist *sg; local
53 sg = areq->src;
54 while (sg) {
55 if ((sg->length % 16) != 0) {
59 if ((sg_dma_len(sg) % 16) != 0) {
63 if (!IS_ALIGNED(sg->offset, 16)) {
67 sg = sg_next(sg);
69 sg = areq->dst;
70 while (sg) {
135 struct scatterlist *sg; local
[all...]
/linux-master/net/core/
H A Dskmsg.c15 if (msg->sg.end > msg->sg.start &&
16 elem_first_coalesce < msg->sg.end)
19 if (msg->sg.end < msg->sg.start &&
20 (elem_first_coalesce > msg->sg.start ||
21 elem_first_coalesce < msg->sg.end))
31 u32 osize = msg->sg.size;
34 len -= msg->sg.size;
52 i = msg->sg
[all...]
/linux-master/drivers/infiniband/hw/mthca/
H A Dmthca_wqe.h117 struct ib_sge *sg)
119 dseg->byte_count = cpu_to_be32(sg->length);
120 dseg->lkey = cpu_to_be32(sg->lkey);
121 dseg->addr = cpu_to_be64(sg->addr);
116 mthca_set_data_seg(struct mthca_data_seg *dseg, struct ib_sge *sg) argument
/linux-master/drivers/gpu/drm/radeon/
H A Dradeon_prime.h38 struct sg_table *sg);
/linux-master/include/linux/
H A Dpage_reporting.h14 struct scatterlist *sg, unsigned int nents);
/linux-master/drivers/scsi/arm/
H A Darm_scsi.h29 * copy_SCp_to_sg() Assumes contiguous allocation at @sg of at-most @max
31 * (possibly chained) sg-list
33 static inline int copy_SCp_to_sg(struct scatterlist *sg, struct scsi_pointer *SCp, int max) argument
42 sg_set_buf(sg, SCp->ptr, SCp->this_residual);
49 *(++sg) = *src_sg;
50 sg_mark_end(sg);
108 struct scatterlist *sg; local
111 scsi_for_each_sg(SCpnt, sg, sg_count, i)
112 len += sg->length;
/linux-master/drivers/dma-buf/
H A Dudmabuf.c29 struct sg_table *sg; member in struct:udmabuf
90 struct sg_table *sg; local
93 sg = kzalloc(sizeof(*sg), GFP_KERNEL);
94 if (!sg)
96 ret = sg_alloc_table_from_pages(sg, ubuf->pages, ubuf->pagecount,
101 ret = dma_map_sgtable(dev, sg, direction, 0);
104 return sg;
107 sg_free_table(sg);
108 kfree(sg);
112 put_sg_table(struct device *dev, struct sg_table *sg, enum dma_data_direction direction) argument
126 unmap_udmabuf(struct dma_buf_attachment *at, struct sg_table *sg, enum dma_data_direction direction) argument
[all...]
/linux-master/drivers/gpu/drm/omapdrm/
H A Domap_gem_dmabuf.c25 struct sg_table *sg; local
26 sg = omap_gem_get_sg(obj, dir);
27 if (IS_ERR(sg))
28 return sg;
30 return sg;
34 struct sg_table *sg, enum dma_data_direction dir)
37 omap_gem_put_sg(obj, sg);
33 omap_gem_unmap_dma_buf(struct dma_buf_attachment *attachment, struct sg_table *sg, enum dma_data_direction dir) argument
/linux-master/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-cipher.c30 struct scatterlist *sg; local
45 sg = areq->src;
46 while (sg) {
47 todo = min(len, sg->length);
52 if (!IS_ALIGNED(sg->offset, 16)) {
57 sg = sg_next(sg);
60 sg = areq->dst;
61 while (sg) {
62 todo = min(len, sg
126 struct scatterlist *sg = areq->src; local
192 struct scatterlist *sg; local
[all...]
/linux-master/drivers/dma/
H A Ddma-axi-dmac.c139 struct axi_dmac_sg sg[] __counted_by(num_sgs);
232 struct axi_dmac_sg *sg; local
251 sg = &desc->sg[desc->num_submitted];
254 if (sg->hw->id != AXI_DMAC_SG_UNUSED) {
255 sg->schedule_when_free = true;
272 sg->hw->id = axi_dmac_read(dmac, AXI_DMAC_REG_TRANSFER_ID);
276 axi_dmac_write(dmac, AXI_DMAC_REG_DEST_ADDRESS, sg->hw->dest_addr);
277 axi_dmac_write(dmac, AXI_DMAC_REG_DEST_STRIDE, sg->hw->dst_stride);
281 axi_dmac_write(dmac, AXI_DMAC_REG_SRC_ADDRESS, sg
318 axi_dmac_total_sg_bytes(struct axi_dmac_chan *chan, struct axi_dmac_sg *sg) argument
331 struct axi_dmac_sg *sg; local
378 struct axi_dmac_sg *sg; local
402 struct axi_dmac_sg *sg; local
582 axi_dmac_fill_linear_sg(struct axi_dmac_chan *chan, enum dma_transfer_direction direction, dma_addr_t addr, unsigned int num_periods, unsigned int period_len, struct axi_dmac_sg *sg) argument
631 struct scatterlist *sg; local
[all...]
/linux-master/drivers/target/tcm_fc/
H A Dtfc_io.c47 struct scatterlist *sg = NULL; local
81 sg = se_cmd->t_data_sg;
82 mem_len = sg->length;
83 mem_off = sg->offset;
84 page = sg_page(sg);
99 sg = sg_next(sg);
100 mem_len = min((size_t)sg->length, remaining);
101 mem_off = sg->offset;
102 page = sg_page(sg);
202 struct scatterlist *sg = NULL; local
[all...]
/linux-master/drivers/crypto/bcm/
H A Dutil.h55 void __dump_sg(struct scatterlist *sg, unsigned int skip, unsigned int len);
57 #define dump_sg(sg, skip, len) __dump_sg(sg, skip, len)
77 static inline void dump_sg(struct scatterlist *sg, unsigned int skip, argument
84 int spu_sg_at_offset(struct scatterlist *sg, unsigned int skip,
87 /* Copy sg data, from skip, length len, to dest */
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c104 static void otx2_dma_unmap_skb_frags(struct otx2_nic *pfvf, struct sg_list *sg) argument
108 for (seg = 0; seg < sg->num_segs; seg++) {
109 otx2_dma_unmap_page(pfvf, sg->dma_addr[seg],
110 sg->size[seg], DMA_TO_DEVICE);
112 sg->num_segs = 0;
120 struct sg_list *sg; local
124 sg = &sq->sg[snd_comp->sqe_id];
126 pa = otx2_iova_to_phys(pfvf->iommu_domain, sg->dma_addr[0]);
127 otx2_dma_unmap_page(pfvf, sg
143 struct sg_list *sg; local
257 struct nix_rx_sg_s *sg = &cqe->sg; local
344 struct nix_rx_sg_s *sg = &cqe->sg; local
617 struct nix_sqe_sg_s *sg = NULL; local
803 struct sg_list *sg = &sq->sg[sqe]; local
835 struct sg_list *sg = &sq->sg[sqe]; local
852 struct nix_sqe_sg_s *sg = NULL; local
1256 struct sg_list *sg; local
1321 struct sg_list *sg; local
1350 struct nix_sqe_sg_s *sg = NULL; local
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_gart.c120 struct sg_table *sg = NULL; local
153 sg = kmalloc(sizeof(*sg), GFP_KERNEL);
154 if (!sg) {
158 ret = sg_alloc_table(sg, 1, GFP_KERNEL);
162 sg_dma_address(sg->sgl) = dma_addr;
163 sg->sgl->length = adev->gart.table_size;
165 sg->sgl->dma_length = adev->gart.table_size;
180 bo->tbo.sg = sg;
229 struct sg_table *sg = adev->gart.bo->tbo.sg; local
[all...]
/linux-master/drivers/crypto/cavium/nitrox/
H A Dnitrox_req.h474 * @sg: crypto request buffer.
482 struct scatterlist *sg; member in struct:nitrox_sgtable
567 * @sg: Destination SG list
573 static inline struct scatterlist *create_single_sg(struct scatterlist *sg, argument
576 sg_set_buf(sg, buf, buflen);
577 sg++;
578 return sg;
582 * create_multi_sg - Create multiple sg entries with buflen data length from
594 struct scatterlist *sg = to_sg; local
602 sg_set_buf(sg, sg_vir
647 struct scatterlist *sg; local
705 struct scatterlist *sg; local
[all...]
/linux-master/kernel/dma/
H A Ddebug.h20 extern void debug_dma_map_sg(struct device *dev, struct scatterlist *sg,
51 struct scatterlist *sg,
55 struct scatterlist *sg,
70 static inline void debug_dma_map_sg(struct device *dev, struct scatterlist *sg, argument
119 struct scatterlist *sg,
125 struct scatterlist *sg,
118 debug_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, int direction) argument
124 debug_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, int direction) argument
/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dqat_bl.c67 struct scatterlist *sg; local
96 for_each_sg(sgl, sg, n, i) {
99 if (!sg->length)
102 if (left >= sg->length) {
103 left -= sg->length;
106 bufl->buffers[y].addr = dma_map_single(dev, sg_virt(sg) + left,
107 sg->length - left,
109 bufl->buffers[y].len = sg->length;
151 for_each_sg(sglout, sg, n_sglout, i) {
154 if (!sg
273 struct scatterlist *sg; local
364 qat_bl_realloc_map_new_dst(struct adf_accel_dev *accel_dev, struct scatterlist **sg, unsigned int dlen, struct qat_request_buffs *qat_bufs, gfp_t gfp) argument
[all...]
/linux-master/kernel/sched/
H A Dtopology.c608 static void free_sched_groups(struct sched_group *sg, int free_sgc) argument
612 if (!sg)
615 first = sg;
617 tmp = sg->next;
619 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref))
620 kfree(sg->sgc);
622 if (atomic_dec_and_test(&sg->ref))
623 kfree(sg);
624 sg = tmp;
625 } while (sg !
761 struct sched_group *sg = sd->groups; local
808 group_balance_cpu(struct sched_group *sg) argument
920 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) argument
959 struct sched_group *sg; local
980 init_overlap_sched_group(struct sched_domain *sd, struct sched_group *sg) argument
1036 struct sched_group *first = NULL, *last = NULL, *sg; local
1203 struct sched_group *sg; local
1259 struct sched_group *sg; local
1292 struct sched_group *sg = sd->groups; local
2235 struct sched_group *sg; local
[all...]
/linux-master/drivers/net/ethernet/microchip/sparx5/
H A Dsparx5_psfp.c132 const struct sparx5_psfp_sg *sg)
138 ips = sparx5_psfp_ipv_to_ips(sg->ipv);
139 base_lsb = sg->basetime.tv_sec & 0xffffffff;
140 base_msb = sg->basetime.tv_sec >> 32;
147 spx5_wr(sg->basetime.tv_nsec, sparx5, ANA_AC_SG_CONFIG_REG_1);
152 ANA_AC_SG_CONFIG_REG_3_LIST_LENGTH_SET(sg->num_entries) |
153 ANA_AC_SG_CONFIG_REG_3_INIT_GATE_STATE_SET(sg->gate_state) |
162 spx5_wr(sg->cycletime, sparx5, ANA_AC_SG_CONFIG_REG_4);
163 spx5_wr(sg->cycletimeext, sparx5, ANA_AC_SG_CONFIG_REG_5);
166 for (i = 0; i < sg
131 sparx5_psfp_sg_set(struct sparx5 *sparx5, u32 id, const struct sparx5_psfp_sg *sg) argument
226 sparx5_psfp_sg_add(struct sparx5 *sparx5, u32 uidx, struct sparx5_psfp_sg *sg, u32 *id) argument
253 const struct sparx5_psfp_sg sg = { 0 }; local
[all...]
/linux-master/arch/mips/jazz/
H A Djazzdma.c547 struct scatterlist *sg; local
549 for_each_sg(sglist, sg, nents, i) {
551 arch_sync_dma_for_device(sg_phys(sg), sg->length,
553 sg->dma_address = vdma_alloc(sg_phys(sg), sg->length);
554 if (sg->dma_address == DMA_MAPPING_ERROR)
556 sg_dma_len(sg) = sg
566 struct scatterlist *sg; local
590 struct scatterlist *sg; local
600 struct scatterlist *sg; local
[all...]
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_pages.c58 obj->mm.page_sizes.sg = 0;
61 obj->mm.page_sizes.sg |= BIT(i);
63 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg));
231 obj->mm.page_sizes.phys = obj->mm.page_sizes.sg = 0;
524 struct scatterlist *sg; local
531 /* As we iterate forward through the sg, we record each entry in a
545 /* We prefer to reuse the last sg so that repeated lookup of this
546 * (or the subsequent) sg are fast - comparing against the last
547 * sg is faster than going through the radixtree.
550 sg
634 struct scatterlist *sg; local
660 struct scatterlist *sg; local
[all...]

Completed in 283 milliseconds

1234567891011>>