Searched refs:segs (Results 1 - 25 of 108) sorted by relevance

12345

/linux-master/include/uapi/linux/
H A Delf-fdpic.h30 struct elf32_fdpic_loadseg segs[]; member in struct:elf32_fdpic_loadmap
45 struct elf64_fdpic_loadseg segs[]; member in struct:elf64_fdpic_loadmap
/linux-master/net/core/
H A Dgso_test.c49 const unsigned int *segs; member in struct:gso_test_case
58 .segs = (const unsigned int[]) { GSO_TEST_SIZE },
65 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
74 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
82 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 },
91 .segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 },
101 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE },
109 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE },
119 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE, 3 },
133 .segs
147 struct sk_buff *skb, *segs, *cur, *next, *last; local
[all...]
H A Dgso.c16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); local
22 segs = ptype->callbacks.gso_segment(skb, features);
28 return segs;
40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); local
53 segs = ptype->callbacks.gso_segment(skb, features);
61 return segs;
91 struct sk_buff *segs; local
124 segs = skb_mac_gso_segment(skb, features);
126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs)))
[all...]
H A Ddev.h115 unsigned int segs)
118 WRITE_ONCE(dev->gso_max_segs, segs);
114 netif_set_gso_max_segs(struct net_device *dev, unsigned int segs) argument
/linux-master/net/mpls/
H A Dmpls_gso.c23 struct sk_buff *segs = ERR_PTR(-EINVAL); local
51 segs = skb_mac_gso_segment(skb, mpls_features);
52 if (IS_ERR_OR_NULL(segs)) {
57 skb = segs;
73 return segs;
/linux-master/net/sctp/
H A Doffload.c42 struct sk_buff *segs = ERR_PTR(-EINVAL); local
68 segs = NULL;
72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG);
73 if (IS_ERR(segs))
78 for (skb = segs; skb; skb = skb->next) {
87 return segs;
/linux-master/net/nsh/
H A Dnsh.c80 struct sk_buff *segs = ERR_PTR(-EINVAL); local
108 segs = skb_mac_gso_segment(skb, features);
109 if (IS_ERR_OR_NULL(segs)) {
115 for (skb = segs; skb; skb = skb->next) {
124 return segs;
/linux-master/net/ipv4/
H A Dtcp_offload.c59 struct sk_buff *segs = ERR_PTR(-EINVAL); local
91 segs = NULL;
100 segs = skb_segment(skb, features);
101 if (IS_ERR(segs))
105 segs->ooo_okay = ooo_okay;
111 if (skb_is_gso(segs))
112 mss *= skb_shinfo(segs)->gso_segs;
116 skb = segs;
121 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss);
178 return segs;
[all...]
H A Dudp_offload.c24 struct sk_buff *segs = ERR_PTR(-EINVAL); local
87 segs = gso_inner_segment(skb, features);
88 if (IS_ERR_OR_NULL(segs)) {
94 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
98 skb = segs;
150 return segs;
160 struct sk_buff *segs = ERR_PTR(-EINVAL); local
182 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment,
188 return segs;
219 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) argument
270 struct sk_buff *segs, *seg; local
379 struct sk_buff *segs = ERR_PTR(-EINVAL); local
[all...]
H A Dgre_offload.c21 struct sk_buff *segs = ERR_PTR(-EINVAL); local
58 segs = skb_mac_gso_segment(skb, features);
59 if (IS_ERR_OR_NULL(segs)) {
65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
69 skb = segs;
119 return segs;
H A Desp4_offload.c129 struct sk_buff *segs = ERR_PTR(-EINVAL); local
135 segs = ops->callbacks.gso_segment(skb, features);
137 return segs;
145 struct sk_buff *segs = ERR_PTR(-EINVAL); local
176 segs = ops->callbacks.gso_segment(skb, features);
178 return segs;
/linux-master/include/linux/
H A Dblk-integrity.h70 unsigned int segs)
72 q->limits.max_integrity_segments = segs;
155 unsigned int segs)
69 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) argument
154 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) argument
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_flow.c624 * @segs: array of one or more packet segments that describe the flow
627 static int ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) argument
633 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK &&
634 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK))
638 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK &&
639 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK))
667 sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ?
671 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4)
673 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6)
675 else if (params->prof->segs[se
1252 ice_flow_find_prof_conds(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, u16 vsi_handle, u32 conds) argument
1346 ice_flow_add_prof_sync(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, struct ice_flow_prof **prof) argument
1537 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, struct ice_flow_prof **prof) argument
1883 ice_flow_set_rss_seg_info(struct ice_flow_seg_info *segs, u8 seg_cnt, const struct ice_rss_hash_cfg *cfg) argument
2234 struct ice_flow_seg_info *segs; local
2381 struct ice_flow_seg_info *segs; local
[all...]
/linux-master/net/ipv6/
H A Dip6_offload.c110 struct sk_buff *segs = ERR_PTR(-EINVAL); local
137 segs = ERR_PTR(-EPROTONOSUPPORT);
152 segs = ops->callbacks.gso_segment(skb, features);
153 if (!segs)
157 if (IS_ERR_OR_NULL(segs))
160 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL);
162 for (skb = segs; skb; skb = skb->next) {
177 kfree_skb_list(segs);
192 return segs;
H A Dudp_offload.c22 struct sk_buff *segs = ERR_PTR(-EINVAL); local
35 segs = skb_udp_tunnel_segment(skb, features, true);
109 segs = skb_segment(skb, features);
113 return segs;
H A Desp6_offload.c165 struct sk_buff *segs = ERR_PTR(-EINVAL); local
171 segs = ops->callbacks.gso_segment(skb, features);
173 return segs;
181 struct sk_buff *segs = ERR_PTR(-EINVAL); local
216 segs = ops->callbacks.gso_segment(skb, features);
218 return segs;
/linux-master/include/rdma/
H A Drdmavt_mr.h28 struct rvt_seg segs[RVT_SEGSZ]; member in struct:rvt_segarray
73 u16 n; /* current index: mr->map[m]->segs[n] */
133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
/linux-master/tools/lib/bpf/
H A Dusdt.c400 static int parse_elf_segs(Elf *elf, const char *path, struct elf_seg **segs, size_t *seg_cnt) argument
427 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs));
431 *segs = tmp;
432 seg = *segs + *seg_cnt;
446 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs);
450 static int parse_vma_segs(int pid, const char *lib_path, struct elf_seg **segs, size_t *seg_cnt) argument
510 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs));
540 find_elf_seg(struct elf_seg *segs, size_t seg_cnt, long virtaddr) argument
556 find_vma_seg(struct elf_seg *segs, size_t seg_cnt, long offset) argument
583 struct elf_seg *segs = NULL, *vma_segs = NULL; local
[all...]
/linux-master/net/sched/
H A Dsch_tbf.c209 struct sk_buff *segs, *nskb; local
214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK);
216 if (IS_ERR_OR_NULL(segs))
220 skb_list_walk_safe(segs, segs, nskb) {
221 skb_mark_not_on_list(segs);
222 qdisc_skb_cb(segs)->pkt_len = segs->len;
223 len += segs->len;
224 ret = qdisc_enqueue(segs,
[all...]
H A Dsch_netem.c424 struct sk_buff *segs; local
427 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK);
429 if (IS_ERR_OR_NULL(segs)) {
434 return segs;
450 struct sk_buff *segs = NULL; local
508 segs = skb->next;
530 /* re-link segs, so that qdisc_drop_all() frees them all */
531 skb->next = segs;
604 if (segs) {
611 while (segs) {
[all...]
/linux-master/drivers/infiniband/sw/rdmavt/
H A Dmr.c379 mr->mr.map[m]->segs[n].vaddr = vaddr;
380 mr->mr.map[m]->segs[n].length = PAGE_SIZE;
569 mr->mr.map[m]->segs[n].vaddr = (void *)addr;
570 mr->mr.map[m]->segs[n].length = ps;
599 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr;
787 while (off >= mr->map[m]->segs[n].length) {
788 off -= mr->map[m]->segs[n].length;
797 isge->vaddr = mr->map[m]->segs[n].vaddr + off;
798 isge->length = mr->map[m]->segs[n].length - off;
894 while (off >= mr->map[m]->segs[
[all...]
/linux-master/fs/
H A Dbinfmt_elf_fdpic.c762 loadmap = kzalloc(struct_size(loadmap, segs, nloads), GFP_KERNEL);
790 seg = loadmap->segs;
815 seg = loadmap->segs;
836 seg = loadmap->segs;
872 mseg = loadmap->segs;
900 seg = loadmap->segs;
933 seg = params->loadmap->segs;
1020 seg = params->loadmap->segs;
1237 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) argument
1255 elf->e_phnum = segs;
1424 fill_extnum_info(struct elfhdr *elf, struct elf_shdr *shdr4extnum, elf_addr_t e_shoff, int segs) argument
1468 int segs; local
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_coredump.h90 u16 segs; member in struct:bnxt_hwrm_dbg_dma_info
H A Dbnxt_coredump.c52 info->segs = le16_to_cpu(*((__le16 *)(resp +
54 if (!info->segs) {
59 info->dest_buf_size = info->segs *
113 coredump->total_segs = info.segs;
/linux-master/include/net/
H A Dudp.h468 struct sk_buff *segs; local
490 segs = __skb_gso_segment(skb, features, false);
491 if (IS_ERR_OR_NULL(segs)) {
501 return segs;

Completed in 423 milliseconds

12345