/linux-master/drivers/accel/qaic/ |
H A D | qaic_data.c | 176 struct scatterlist *sg, *sgn, *sgf, *sgl; local 187 for (sg = sgt_in->sgl; sg; sg = sg_next(sg)) { 188 len = sg_dma_len(sg); 193 sgf = sg; 200 sgl = sg; 222 /* copy relevant sg node and fix page and length */ 224 for_each_sgtable_sg(sgt, sg, 257 struct scatterlist *sg; local 442 struct scatterlist *sg; local 575 struct scatterlist *sg; local 604 struct scatterlist *sg; local [all...] |
/linux-master/security/integrity/ima/ |
H A D | ima_crypto.c | 332 struct scatterlist sg[1]; local 412 sg_init_one(&sg[0], rbuf[active], rbuf_len); 413 ahash_request_set_crypt(req, sg, NULL, rbuf_len); 663 struct scatterlist sg; local 682 sg_init_one(&sg, buf, len); 683 ahash_request_set_crypt(req, &sg, NULL, len);
|
/linux-master/net/tls/ |
H A D | tls_sw.c | 68 struct scatterlist sg[]; member in struct:tls_decrypt_ctx 195 struct scatterlist *sg; local 230 for_each_sg(sg_next(sgout), sg, UINT_MAX, pages) { local 231 if (!sg) 233 put_page(sg_page(sg)); 340 /* We add page references worth len bytes from encrypted sg 341 * at the end of plaintext sg. It is guaranteed that msg_en 344 len = required - msg_pl->sg.size; 349 skip = prot->prepend_size + msg_pl->sg.size; 446 &msg_en->sg [all...] |
H A D | tls_device.c | 888 struct scatterlist sg[1]; local 910 sg_init_table(sg, 1); 911 sg_set_buf(&sg[0], buf, 918 err = decrypt_skb(sk, sg);
|
H A D | tls_device_fallback.c | 40 static void chain_to_walk(struct scatterlist *sg, struct scatter_walk *walk) argument 42 struct scatterlist *src = walk->sg; 45 sg_set_page(sg, sg_page(src), 48 scatterwalk_crypto_chain(sg, sg_next(src), 2);
|
/linux-master/net/smc/ |
H A D | smc_core.c | 2078 struct scatterlist *sg; local 2099 for_each_sg(buf_desc->sgt[lnk->link_idx].sgl, sg, nents, i) { 2101 sg_set_page(sg, vmalloc_to_page(buf), size, offset); 2112 /* map sg table to DMA address */
|
/linux-master/net/ipv6/ |
H A D | tcp_ipv6.c | 700 struct scatterlist sg; local 714 sg_init_one(&sg, bp, sizeof(*bp) + sizeof(*th)); 715 ahash_request_set_crypt(hp->req, &sg, NULL,
|
H A D | esp6.c | 122 struct scatterlist *sg; local 131 * Skip the first sg which comes from skb->data. 134 for (sg = sg_next(req->src); sg; sg = sg_next(sg)) 135 skb_page_unref(sg_page(sg), skb->pp_recycle); 574 struct scatterlist *sg, *dsg; local 597 sg = esp_req_sg(aead, req); 600 dsg = sg; 928 struct scatterlist *sg; local [all...] |
/linux-master/net/ipv4/ |
H A D | tcp_ipv4.c | 1568 struct scatterlist sg; local 1582 sg_init_one(&sg, bp, sizeof(*bp) + sizeof(*th)); 1583 ahash_request_set_crypt(hp->req, &sg, NULL,
|
H A D | tcp.c | 4444 struct scatterlist sg; local 4446 sg_init_one(&sg, key->key, keylen); 4447 ahash_request_set_crypt(hp->req, &sg, NULL, keylen);
|
H A D | esp4.c | 105 struct scatterlist *sg; local 114 * Skip the first sg which comes from skb->data. 117 for (sg = sg_next(req->src); sg; sg = sg_next(sg)) 118 skb_page_unref(sg_page(sg), skb->pp_recycle); 537 struct scatterlist *sg, *dsg; local 560 sg = esp_req_sg(aead, req); 563 dsg = sg; 883 struct scatterlist *sg; local [all...] |
/linux-master/net/core/ |
H A D | skbuff.c | 4605 bool csum, sg; local 4636 sg = !!(features & NETIF_F_SG); 4639 if (sg && csum && (mss != GSO_BY_FRAGS)) { 4710 (skb_headlen(list_skb) == len || sg)) { 4754 if (hsize > len || !sg) 4786 if (!sg) { 5030 __skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len, argument 5044 sg_set_buf(sg, skb->data + offset, copy); 5059 if (unlikely(elt && sg_is_last(&sg[elt - 1]))) 5064 sg_set_page(&sg[el 5113 skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len) argument 5145 skb_to_sgvec_nomark(struct sk_buff *skb, struct scatterlist *sg, int offset, int len) argument [all...] |
H A D | filter.c | 2612 u32 i = msg->sg.start; 2618 if (len >= msg->sg.size) 2620 } while (i != msg->sg.end); 2622 msg->sg.curr = i; 2623 msg->sg.copybreak = 0; 2647 i = msg->sg.start; 2654 } while (i != msg->sg.end); 2660 /* The start may point into the sg element so we need to also 2664 if (!test_bit(i, msg->sg.copy) && bytes_sg_total <= len) 2675 * will copy the entire sg entr [all...] |
H A D | datagram.c | 479 struct scatterlist sg; local 483 sg_init_one(&sg, addr, copied); 484 ahash_request_set_crypt(hash, &sg, NULL, copied);
|
/linux-master/kernel/sched/ |
H A D | topology.c | 608 static void free_sched_groups(struct sched_group *sg, int free_sgc) argument 612 if (!sg) 615 first = sg; 617 tmp = sg->next; 619 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref)) 620 kfree(sg->sgc); 622 if (atomic_dec_and_test(&sg->ref)) 623 kfree(sg); 624 sg = tmp; 625 } while (sg ! 761 struct sched_group *sg = sd->groups; local 808 group_balance_cpu(struct sched_group *sg) argument 920 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) argument 959 struct sched_group *sg; local 980 init_overlap_sched_group(struct sched_domain *sd, struct sched_group *sg) argument 1036 struct sched_group *first = NULL, *last = NULL, *sg; local 1203 struct sched_group *sg; local 1259 struct sched_group *sg; local 1292 struct sched_group *sg = sd->groups; local 2242 struct sched_group *sg; local [all...] |
H A D | sched.h | 1251 static inline struct cpumask *sched_group_span(struct sched_group *sg); 1944 static inline struct cpumask *sched_group_span(struct sched_group *sg) argument 1946 return to_cpumask(sg->cpumask); 1952 static inline struct cpumask *group_balance_mask(struct sched_group *sg) argument 1954 return to_cpumask(sg->sgc->cpumask); 1957 extern int group_balance_cpu(struct sched_group *sg);
|
H A D | fair.c | 7426 struct sched_group *sg = sd->groups; local 7428 if (sg->flags & SD_CLUSTER) { 7429 for_each_cpu_wrap(cpu, sched_group_span(sg), target + 1) { 7445 cpumask_andnot(cpus, cpus, sched_group_span(sg)); 10029 * @sg: sched_group candidate to be checked for being the busiest 10032 * Determine if @sg is a busier group than the previously selected 10035 * Return: %true if @sg is a busier group than the previously selected 10040 struct sched_group *sg, 10057 (!capacity_greater(capacity_of(env->dst_cpu), sg->sgc->max_capacity) || 10086 return sched_asym_prefer(sds->busiest->asym_prefer_cpu, sg 10038 update_sd_pick_busiest(struct lb_env *env, struct sd_lb_stats *sds, struct sched_group *sg, struct sg_lb_stats *sgs) argument 10617 struct sched_group *sg = env->sd->groups; local 11237 struct sched_group *sg = env->sd->groups; local [all...] |
/linux-master/include/scsi/ |
H A D | scsi_cmnd.h | 166 extern void *scsi_kmap_atomic_sg(struct scatterlist *sg, int sg_count, 206 #define scsi_for_each_sg(cmd, sg, nseg, __i) \ 207 for_each_sg(scsi_sglist(cmd), sg, nseg, __i) 331 #define scsi_for_each_prot_sg(cmd, sg, nseg, __i) \ 332 for_each_sg(scsi_prot_sglist(cmd), sg, nseg, __i)
|
/linux-master/include/linux/ |
H A D | virtio.h | 28 * sg element. 43 struct scatterlist sg[], unsigned int num, 48 struct scatterlist sg[], unsigned int num, 53 struct scatterlist sg[], unsigned int num,
|
H A D | skmsg.h | 44 struct sk_msg_sg sg; member in struct:sk_msg 146 WARN_ON(i == msg->sg.end && bytes); 180 sk_msg_iter_var_prev(msg->sg.which) 183 sk_msg_iter_var_next(msg->sg.which) 187 BUILD_BUG_ON(ARRAY_SIZE(msg->sg.data) - 1 != NR_MSG_FRAG_IDS); 189 sg_init_marker(msg->sg.data, NR_MSG_FRAG_IDS); 195 dst->sg.data[which] = src->sg.data[which]; 196 dst->sg.data[which].length = size; 197 dst->sg [all...] |
H A D | skbuff.h | 1381 int __must_check skb_to_sgvec_nomark(struct sk_buff *skb, struct scatterlist *sg, 1383 int __must_check skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg,
|
H A D | libata.h | 621 struct scatterlist *sg; member in struct:ata_queued_cmd 1135 extern void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg, 1831 qc->sg = NULL;
|
/linux-master/include/linux/sched/ |
H A D | topology.h | 191 struct sched_group *__percpu *sg; member in struct:sd_data
|
/linux-master/include/crypto/ |
H A D | algapi.h | 122 struct scatterlist *sg; member in struct:scatter_walk
|
/linux-master/fs/smb/client/ |
H A D | smb2ops.c | 4228 struct scatterlist *sg; local 4271 creq = smb2_get_aead_req(tfm, rqst, num_rqst, sign, &iv, &req, &sg, 4290 aead_request_set_crypt(req, sg, sg, crypt_len, iv);
|