Searched refs:sg (Results 51 - 75 of 900) sorted by last modified time

1234567891011>>

/linux-master/drivers/accel/qaic/
H A Dqaic_data.c176 struct scatterlist *sg, *sgn, *sgf, *sgl; local
187 for (sg = sgt_in->sgl; sg; sg = sg_next(sg)) {
188 len = sg_dma_len(sg);
193 sgf = sg;
200 sgl = sg;
222 /* copy relevant sg node and fix page and length */
224 for_each_sgtable_sg(sgt, sg,
257 struct scatterlist *sg; local
442 struct scatterlist *sg; local
575 struct scatterlist *sg; local
604 struct scatterlist *sg; local
[all...]
/linux-master/security/integrity/ima/
H A Dima_crypto.c332 struct scatterlist sg[1]; local
412 sg_init_one(&sg[0], rbuf[active], rbuf_len);
413 ahash_request_set_crypt(req, sg, NULL, rbuf_len);
663 struct scatterlist sg; local
682 sg_init_one(&sg, buf, len);
683 ahash_request_set_crypt(req, &sg, NULL, len);
/linux-master/net/tls/
H A Dtls_sw.c68 struct scatterlist sg[]; member in struct:tls_decrypt_ctx
195 struct scatterlist *sg; local
230 for_each_sg(sg_next(sgout), sg, UINT_MAX, pages) { local
231 if (!sg)
233 put_page(sg_page(sg));
340 /* We add page references worth len bytes from encrypted sg
341 * at the end of plaintext sg. It is guaranteed that msg_en
344 len = required - msg_pl->sg.size;
349 skip = prot->prepend_size + msg_pl->sg.size;
446 &msg_en->sg
[all...]
H A Dtls_device.c888 struct scatterlist sg[1]; local
910 sg_init_table(sg, 1);
911 sg_set_buf(&sg[0], buf,
918 err = decrypt_skb(sk, sg);
H A Dtls_device_fallback.c40 static void chain_to_walk(struct scatterlist *sg, struct scatter_walk *walk) argument
42 struct scatterlist *src = walk->sg;
45 sg_set_page(sg, sg_page(src),
48 scatterwalk_crypto_chain(sg, sg_next(src), 2);
/linux-master/net/smc/
H A Dsmc_core.c2078 struct scatterlist *sg; local
2099 for_each_sg(buf_desc->sgt[lnk->link_idx].sgl, sg, nents, i) {
2101 sg_set_page(sg, vmalloc_to_page(buf), size, offset);
2112 /* map sg table to DMA address */
/linux-master/net/ipv6/
H A Dtcp_ipv6.c700 struct scatterlist sg; local
714 sg_init_one(&sg, bp, sizeof(*bp) + sizeof(*th));
715 ahash_request_set_crypt(hp->req, &sg, NULL,
H A Desp6.c122 struct scatterlist *sg; local
131 * Skip the first sg which comes from skb->data.
134 for (sg = sg_next(req->src); sg; sg = sg_next(sg))
135 skb_page_unref(sg_page(sg), skb->pp_recycle);
574 struct scatterlist *sg, *dsg; local
597 sg = esp_req_sg(aead, req);
600 dsg = sg;
928 struct scatterlist *sg; local
[all...]
/linux-master/net/ipv4/
H A Dtcp_ipv4.c1568 struct scatterlist sg; local
1582 sg_init_one(&sg, bp, sizeof(*bp) + sizeof(*th));
1583 ahash_request_set_crypt(hp->req, &sg, NULL,
H A Dtcp.c4444 struct scatterlist sg; local
4446 sg_init_one(&sg, key->key, keylen);
4447 ahash_request_set_crypt(hp->req, &sg, NULL, keylen);
H A Desp4.c105 struct scatterlist *sg; local
114 * Skip the first sg which comes from skb->data.
117 for (sg = sg_next(req->src); sg; sg = sg_next(sg))
118 skb_page_unref(sg_page(sg), skb->pp_recycle);
537 struct scatterlist *sg, *dsg; local
560 sg = esp_req_sg(aead, req);
563 dsg = sg;
883 struct scatterlist *sg; local
[all...]
/linux-master/net/core/
H A Dskbuff.c4605 bool csum, sg; local
4636 sg = !!(features & NETIF_F_SG);
4639 if (sg && csum && (mss != GSO_BY_FRAGS)) {
4710 (skb_headlen(list_skb) == len || sg)) {
4754 if (hsize > len || !sg)
4786 if (!sg) {
5030 __skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len, argument
5044 sg_set_buf(sg, skb->data + offset, copy);
5059 if (unlikely(elt && sg_is_last(&sg[elt - 1])))
5064 sg_set_page(&sg[el
5113 skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len) argument
5145 skb_to_sgvec_nomark(struct sk_buff *skb, struct scatterlist *sg, int offset, int len) argument
[all...]
H A Dfilter.c2612 u32 i = msg->sg.start;
2618 if (len >= msg->sg.size)
2620 } while (i != msg->sg.end);
2622 msg->sg.curr = i;
2623 msg->sg.copybreak = 0;
2647 i = msg->sg.start;
2654 } while (i != msg->sg.end);
2660 /* The start may point into the sg element so we need to also
2664 if (!test_bit(i, msg->sg.copy) && bytes_sg_total <= len)
2675 * will copy the entire sg entr
[all...]
H A Ddatagram.c479 struct scatterlist sg; local
483 sg_init_one(&sg, addr, copied);
484 ahash_request_set_crypt(hash, &sg, NULL, copied);
/linux-master/kernel/sched/
H A Dtopology.c608 static void free_sched_groups(struct sched_group *sg, int free_sgc) argument
612 if (!sg)
615 first = sg;
617 tmp = sg->next;
619 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref))
620 kfree(sg->sgc);
622 if (atomic_dec_and_test(&sg->ref))
623 kfree(sg);
624 sg = tmp;
625 } while (sg !
761 struct sched_group *sg = sd->groups; local
808 group_balance_cpu(struct sched_group *sg) argument
920 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) argument
959 struct sched_group *sg; local
980 init_overlap_sched_group(struct sched_domain *sd, struct sched_group *sg) argument
1036 struct sched_group *first = NULL, *last = NULL, *sg; local
1203 struct sched_group *sg; local
1259 struct sched_group *sg; local
1292 struct sched_group *sg = sd->groups; local
2242 struct sched_group *sg; local
[all...]
H A Dsched.h1251 static inline struct cpumask *sched_group_span(struct sched_group *sg);
1944 static inline struct cpumask *sched_group_span(struct sched_group *sg) argument
1946 return to_cpumask(sg->cpumask);
1952 static inline struct cpumask *group_balance_mask(struct sched_group *sg) argument
1954 return to_cpumask(sg->sgc->cpumask);
1957 extern int group_balance_cpu(struct sched_group *sg);
H A Dfair.c7426 struct sched_group *sg = sd->groups; local
7428 if (sg->flags & SD_CLUSTER) {
7429 for_each_cpu_wrap(cpu, sched_group_span(sg), target + 1) {
7445 cpumask_andnot(cpus, cpus, sched_group_span(sg));
10029 * @sg: sched_group candidate to be checked for being the busiest
10032 * Determine if @sg is a busier group than the previously selected
10035 * Return: %true if @sg is a busier group than the previously selected
10040 struct sched_group *sg,
10057 (!capacity_greater(capacity_of(env->dst_cpu), sg->sgc->max_capacity) ||
10086 return sched_asym_prefer(sds->busiest->asym_prefer_cpu, sg
10038 update_sd_pick_busiest(struct lb_env *env, struct sd_lb_stats *sds, struct sched_group *sg, struct sg_lb_stats *sgs) argument
10617 struct sched_group *sg = env->sd->groups; local
11237 struct sched_group *sg = env->sd->groups; local
[all...]
/linux-master/include/scsi/
H A Dscsi_cmnd.h166 extern void *scsi_kmap_atomic_sg(struct scatterlist *sg, int sg_count,
206 #define scsi_for_each_sg(cmd, sg, nseg, __i) \
207 for_each_sg(scsi_sglist(cmd), sg, nseg, __i)
331 #define scsi_for_each_prot_sg(cmd, sg, nseg, __i) \
332 for_each_sg(scsi_prot_sglist(cmd), sg, nseg, __i)
/linux-master/include/linux/
H A Dvirtio.h28 * sg element.
43 struct scatterlist sg[], unsigned int num,
48 struct scatterlist sg[], unsigned int num,
53 struct scatterlist sg[], unsigned int num,
H A Dskmsg.h44 struct sk_msg_sg sg; member in struct:sk_msg
146 WARN_ON(i == msg->sg.end && bytes);
180 sk_msg_iter_var_prev(msg->sg.which)
183 sk_msg_iter_var_next(msg->sg.which)
187 BUILD_BUG_ON(ARRAY_SIZE(msg->sg.data) - 1 != NR_MSG_FRAG_IDS);
189 sg_init_marker(msg->sg.data, NR_MSG_FRAG_IDS);
195 dst->sg.data[which] = src->sg.data[which];
196 dst->sg.data[which].length = size;
197 dst->sg
[all...]
H A Dskbuff.h1381 int __must_check skb_to_sgvec_nomark(struct sk_buff *skb, struct scatterlist *sg,
1383 int __must_check skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg,
H A Dlibata.h621 struct scatterlist *sg; member in struct:ata_queued_cmd
1135 extern void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg,
1831 qc->sg = NULL;
/linux-master/include/linux/sched/
H A Dtopology.h191 struct sched_group *__percpu *sg; member in struct:sd_data
/linux-master/include/crypto/
H A Dalgapi.h122 struct scatterlist *sg; member in struct:scatter_walk
/linux-master/fs/smb/client/
H A Dsmb2ops.c4228 struct scatterlist *sg; local
4271 creq = smb2_get_aead_req(tfm, rqst, num_rqst, sign, &iv, &req, &sg,
4290 aead_request_set_crypt(req, sg, sg, crypt_len, iv);

Completed in 418 milliseconds

1234567891011>>