/linux-master/include/uapi/linux/ |
H A D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member in struct:elf32_fdpic_loadmap 45 struct elf64_fdpic_loadseg segs[]; member in struct:elf64_fdpic_loadmap
|
/linux-master/net/core/ |
H A D | gso_test.c | 49 const unsigned int *segs; member in struct:gso_test_case 58 .segs = (const unsigned int[]) { GSO_TEST_SIZE }, 65 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 }, 74 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 }, 82 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 }, 91 .segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 }, 101 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE }, 109 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE }, 119 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE, 3 }, 133 .segs 147 struct sk_buff *skb, *segs, *cur, *next, *last; local [all...] |
H A D | gso.c | 16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); local 22 segs = ptype->callbacks.gso_segment(skb, features); 28 return segs; 40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); local 53 segs = ptype->callbacks.gso_segment(skb, features); 61 return segs; 91 struct sk_buff *segs; local 124 segs = skb_mac_gso_segment(skb, features); 126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) [all...] |
H A D | dev.h | 115 unsigned int segs) 118 WRITE_ONCE(dev->gso_max_segs, segs); 114 netif_set_gso_max_segs(struct net_device *dev, unsigned int segs) argument
|
/linux-master/net/mpls/ |
H A D | mpls_gso.c | 23 struct sk_buff *segs = ERR_PTR(-EINVAL); local 51 segs = skb_mac_gso_segment(skb, mpls_features); 52 if (IS_ERR_OR_NULL(segs)) { 57 skb = segs; 73 return segs;
|
/linux-master/net/sctp/ |
H A D | offload.c | 42 struct sk_buff *segs = ERR_PTR(-EINVAL); local 68 segs = NULL; 72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); 73 if (IS_ERR(segs)) 78 for (skb = segs; skb; skb = skb->next) { 87 return segs;
|
/linux-master/net/nsh/ |
H A D | nsh.c | 80 struct sk_buff *segs = ERR_PTR(-EINVAL); local 108 segs = skb_mac_gso_segment(skb, features); 109 if (IS_ERR_OR_NULL(segs)) { 115 for (skb = segs; skb; skb = skb->next) { 124 return segs;
|
/linux-master/net/ipv4/ |
H A D | tcp_offload.c | 59 struct sk_buff *segs = ERR_PTR(-EINVAL); local 91 segs = NULL; 100 segs = skb_segment(skb, features); 101 if (IS_ERR(segs)) 105 segs->ooo_okay = ooo_okay; 111 if (skb_is_gso(segs)) 112 mss *= skb_shinfo(segs)->gso_segs; 116 skb = segs; 121 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); 178 return segs; [all...] |
H A D | udp_offload.c | 24 struct sk_buff *segs = ERR_PTR(-EINVAL); local 87 segs = gso_inner_segment(skb, features); 88 if (IS_ERR_OR_NULL(segs)) { 94 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); 98 skb = segs; 150 return segs; 160 struct sk_buff *segs = ERR_PTR(-EINVAL); local 182 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, 188 return segs; 219 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) argument 270 struct sk_buff *segs, *seg; local 379 struct sk_buff *segs = ERR_PTR(-EINVAL); local [all...] |
H A D | gre_offload.c | 21 struct sk_buff *segs = ERR_PTR(-EINVAL); local 58 segs = skb_mac_gso_segment(skb, features); 59 if (IS_ERR_OR_NULL(segs)) { 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); 69 skb = segs; 119 return segs;
|
H A D | esp4_offload.c | 129 struct sk_buff *segs = ERR_PTR(-EINVAL); local 135 segs = ops->callbacks.gso_segment(skb, features); 137 return segs; 145 struct sk_buff *segs = ERR_PTR(-EINVAL); local 176 segs = ops->callbacks.gso_segment(skb, features); 178 return segs;
|
/linux-master/include/linux/ |
H A D | blk-integrity.h | 70 unsigned int segs) 72 q->limits.max_integrity_segments = segs; 155 unsigned int segs) 69 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) argument 154 blk_queue_max_integrity_segments(struct request_queue *q, unsigned int segs) argument
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_flow.c | 624 * @segs: array of one or more packet segments that describe the flow 627 static int ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) argument 633 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && 634 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) 638 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && 639 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) 667 sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ? 671 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) 673 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) 675 else if (params->prof->segs[se 1252 ice_flow_find_prof_conds(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, u16 vsi_handle, u32 conds) argument 1346 ice_flow_add_prof_sync(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, struct ice_flow_prof **prof) argument 1537 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, struct ice_flow_seg_info *segs, u8 segs_cnt, bool symm, struct ice_flow_prof **prof) argument 1883 ice_flow_set_rss_seg_info(struct ice_flow_seg_info *segs, u8 seg_cnt, const struct ice_rss_hash_cfg *cfg) argument 2234 struct ice_flow_seg_info *segs; local 2381 struct ice_flow_seg_info *segs; local [all...] |
/linux-master/net/ipv6/ |
H A D | ip6_offload.c | 110 struct sk_buff *segs = ERR_PTR(-EINVAL); local 137 segs = ERR_PTR(-EPROTONOSUPPORT); 152 segs = ops->callbacks.gso_segment(skb, features); 153 if (!segs) 157 if (IS_ERR_OR_NULL(segs)) 160 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); 162 for (skb = segs; skb; skb = skb->next) { 177 kfree_skb_list(segs); 192 return segs;
|
H A D | udp_offload.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); local 35 segs = skb_udp_tunnel_segment(skb, features, true); 109 segs = skb_segment(skb, features); 113 return segs;
|
H A D | esp6_offload.c | 165 struct sk_buff *segs = ERR_PTR(-EINVAL); local 171 segs = ops->callbacks.gso_segment(skb, features); 173 return segs; 181 struct sk_buff *segs = ERR_PTR(-EINVAL); local 216 segs = ops->callbacks.gso_segment(skb, features); 218 return segs;
|
/linux-master/include/rdma/ |
H A D | rdmavt_mr.h | 28 struct rvt_seg segs[RVT_SEGSZ]; member in struct:rvt_segarray 73 u16 n; /* current index: mr->map[m]->segs[n] */ 133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; 134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
|
/linux-master/tools/lib/bpf/ |
H A D | usdt.c | 400 static int parse_elf_segs(Elf *elf, const char *path, struct elf_seg **segs, size_t *seg_cnt) argument 427 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); 431 *segs = tmp; 432 seg = *segs + *seg_cnt; 446 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); 450 static int parse_vma_segs(int pid, const char *lib_path, struct elf_seg **segs, size_t *seg_cnt) argument 510 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); 540 find_elf_seg(struct elf_seg *segs, size_t seg_cnt, long virtaddr) argument 556 find_vma_seg(struct elf_seg *segs, size_t seg_cnt, long offset) argument 583 struct elf_seg *segs = NULL, *vma_segs = NULL; local [all...] |
/linux-master/net/sched/ |
H A D | sch_tbf.c | 209 struct sk_buff *segs, *nskb; local 214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); 216 if (IS_ERR_OR_NULL(segs)) 220 skb_list_walk_safe(segs, segs, nskb) { 221 skb_mark_not_on_list(segs); 222 qdisc_skb_cb(segs)->pkt_len = segs->len; 223 len += segs->len; 224 ret = qdisc_enqueue(segs, [all...] |
H A D | sch_netem.c | 424 struct sk_buff *segs; local 427 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); 429 if (IS_ERR_OR_NULL(segs)) { 434 return segs; 450 struct sk_buff *segs = NULL; local 508 segs = skb->next; 530 /* re-link segs, so that qdisc_drop_all() frees them all */ 531 skb->next = segs; 604 if (segs) { 611 while (segs) { [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | mr.c | 379 mr->mr.map[m]->segs[n].vaddr = vaddr; 380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; 569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; 570 mr->mr.map[m]->segs[n].length = ps; 599 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; 787 while (off >= mr->map[m]->segs[n].length) { 788 off -= mr->map[m]->segs[n].length; 797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; 798 isge->length = mr->map[m]->segs[n].length - off; 894 while (off >= mr->map[m]->segs[ [all...] |
/linux-master/fs/ |
H A D | binfmt_elf_fdpic.c | 762 loadmap = kzalloc(struct_size(loadmap, segs, nloads), GFP_KERNEL); 790 seg = loadmap->segs; 815 seg = loadmap->segs; 836 seg = loadmap->segs; 872 mseg = loadmap->segs; 900 seg = loadmap->segs; 933 seg = params->loadmap->segs; 1020 seg = params->loadmap->segs; 1237 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) argument 1255 elf->e_phnum = segs; 1424 fill_extnum_info(struct elfhdr *elf, struct elf_shdr *shdr4extnum, elf_addr_t e_shoff, int segs) argument 1468 int segs; local [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_coredump.h | 90 u16 segs; member in struct:bnxt_hwrm_dbg_dma_info
|
H A D | bnxt_coredump.c | 52 info->segs = le16_to_cpu(*((__le16 *)(resp + 54 if (!info->segs) { 59 info->dest_buf_size = info->segs * 113 coredump->total_segs = info.segs;
|
/linux-master/include/net/ |
H A D | udp.h | 468 struct sk_buff *segs; local 490 segs = __skb_gso_segment(skb, features, false); 491 if (IS_ERR_OR_NULL(segs)) { 501 return segs;
|