/linux-master/include/linux/ |
H A D | maple_tree.h | 74 * This metadata is used to optimize the gap updating code and in reverse 79 unsigned char gap; member in struct:maple_metadata 128 unsigned long gap[MAPLE_ARANGE64_SLOTS]; member in struct:maple_arange_64
|
H A D | mm.h | 3549 unsigned long gap = stack_guard_start_gap(vma); local 3552 vm_start -= gap;
|
H A D | nfs_fs.h | 212 * might fill in the gaps, or possibly a gap is caused 228 } gap[16]; member in struct:nfs_inode::__anon855
|
/linux-master/include/uapi/linux/ |
H A D | fd.h | 30 unsigned char gap, /* gap1 size */ member in struct:floppy_struct
|
H A D | pkt_sched.h | 507 __u32 gap; /* re-ordering gap (0 for none) */ member in struct:tc_netem_qopt
|
/linux-master/lib/ |
H A D | maple_tree.c | 129 unsigned long gap[MAPLE_BIG_NODE_GAPS]; member in struct:maple_big_node::__anon22::__anon23 678 return node->ma64.gap; 866 * @offset: The offset of the highest sub-gap in this node. 874 meta->gap = offset; 883 * @offset: The offset of the highest sub-gap in this node. 913 meta->gap = 0; 931 * ma_meta_gap() - Get the largest gap location of a node from the metadata 936 return mn->ma64.meta.gap; 940 * ma_set_meta_gap() - Set the largest gap location in a nodes metadata 943 * @offset: The location of the largest gap 1030 mte_set_gap(const struct maple_enode *mn, unsigned char gap, unsigned long val) argument 1472 unsigned long pstart, gap, max_gap; local 3022 unsigned long *l_pivs, *pivs, gap; local 3977 bool gap = false; local 4749 unsigned long gap = 0; local 4832 unsigned long pivot, min, gap = 0; local 7296 unsigned long gap = 0, max_gap = 0; local [all...] |
H A D | test_maple_tree.c | 777 pr_debug("Search from %lu-%lu for gap %lu should be at %lu\n", 786 pr_debug("gap %lu %lu\n", (holes[i] >> 12), 900 /* Test filling entire gap. */ 1555 * At this point, there is a gap of 2 at index + 1 between seq100[3] and 1556 * seq100[4]. Search for the gap. 1586 * At this point, there is a gap of 3 at seq100[6]. Find it by 1613 * gap. 1804 int gap = 4; local 1807 gap = 3; 1808 mtree_store_range(mt, i, i + gap, xa_mk_valu 1834 int gap = 4; local 2640 check_dup_gaps(struct maple_tree *mt, unsigned long nr_entries, bool zero_start, unsigned long gap) argument [all...] |
/linux-master/mm/damon/ |
H A D | vaddr.c | 130 unsigned long gap; local 136 gap = vma->vm_start - prev->vm_end; 138 if (gap > sz_range(&first_gap)) { 142 } else if (gap > sz_range(&second_gap)) { 219 * As usual memory map of processes is as below, the gap between the heap and 220 * the uppermost mmap()-ed region, and the gap between the lowermost mmap()-ed
|
/linux-master/mm/ |
H A D | mmap.c | 1578 unsigned long length, gap; local 1597 * Adjust for the gap first so it doesn't interfere with the 1599 * fulill the start gap, the next steps is the minimum to align 1602 gap = vma_iter_addr(&vmi) + info->start_gap; 1603 gap += (info->align_offset - gap) & info->align_mask; 1606 if (vm_start_gap(tmp) < gap + length - 1) { 1613 if (tmp && vm_end_gap(tmp) > gap) { 1620 return gap; 1635 unsigned long length, gap, gap_en local [all...] |
H A D | util.c | 443 unsigned long gap = rlim_stack->rlim_cur; 451 if (gap + pad > gap) 452 gap += pad; 454 if (gap < MIN_GAP) 455 gap = MIN_GAP; 456 else if (gap > MAX_GAP) 457 gap = MAX_GAP; 459 return PAGE_ALIGN(STACK_TOP - gap - rnd);
|
/linux-master/net/bridge/netfilter/ |
H A D | ebtables.c | 689 size_t gap; local 747 gap = e->next_offset - e->target_offset; 764 if (gap < sizeof(struct ebt_standard_target)) { 773 } else if (t->target_size > gap - sizeof(struct ebt_entry_target)) {
|
/linux-master/net/netfilter/ipvs/ |
H A D | ip_vs_est.c | 177 long gap; local 203 gap = kd->est_timer - now; 204 if (gap > 0) { 205 if (gap > IPVS_EST_TICK) { 207 gap = IPVS_EST_TICK; 209 schedule_timeout(gap); 212 if (gap < -8 * IPVS_EST_TICK) 248 long gap; local 255 gap = kd->est_timer - now; 257 if (abs(gap) > [all...] |
/linux-master/net/sched/ |
H A D | sch_netem.c | 93 u32 gap; member in struct:netem_sched_data 217 * TX_IN_GAP_PERIOD => successfully transmitted packets within a gap period 218 * LOST_IN_GAP_PERIOD => isolated losses within a gap period 539 if (q->gap == 0 || /* not doing reordering */ 540 q->counter < q->gap - 1 || /* inside last reordering gap */ 1010 q->gap = qopt->gap; 1016 * if gap is set, need to assume 100% probability 1018 if (q->gap) [all...] |
/linux-master/net/sctp/ |
H A D | tsnmap.c | 71 u32 gap; local 84 gap = tsn - map->base_tsn; 87 if (gap < map->len && test_bit(gap, map->tsn_map)) 98 u16 gap; local 103 gap = tsn - map->base_tsn; 105 if (gap >= map->len && !sctp_tsnmap_grow(map, gap + 1)) 108 if (!sctp_tsnmap_has_gap(map) && gap == 0) { 120 /* Either we already have a gap, o 192 u32 gap; local 253 u32 gap; local 302 u32 gap; local [all...] |
/linux-master/net/tipc/ |
H A D | link.c | 130 * @last_gap: last gap ack blocks for bcast (FIXME) 131 * @last_ga: ptr to gap ack blocks 266 u16 acked, u16 gap, 411 u16 gap = 0; local 414 gap = l->snd_nxt - l->rcv_nxt; 416 gap = buf_seqno(skb) - l->rcv_nxt; 417 return gap; 1470 gacks[n].gap = htons(seqno - expect); 1486 gacks[n].gap = 0; 1532 * @gap 1541 tipc_link_advance_transmq(struct tipc_link *l, struct tipc_link *r, u16 acked, u16 gap, struct tipc_gap_ack_blks *ga, struct sk_buff_head *xmitq, bool *retransmitted, int *rc) argument 2192 u16 gap = msg_seq_gap(hdr); local 2469 tipc_link_bc_ack_rcv(struct tipc_link *r, u16 acked, u16 gap, struct tipc_gap_ack_blks *ga, struct sk_buff_head *xmitq, struct sk_buff_head *retrq) argument [all...] |
H A D | link.h | 147 int tipc_link_bc_ack_rcv(struct tipc_link *l, u16 acked, u16 gap,
|
H A D | msg.h | 154 * @gap: number of gap packets since the last ack 162 __be16 gap; member in struct:tipc_gap_ack 177 * | gap | ack | | 181 * | gap | ack | |
|
/linux-master/tools/include/uapi/linux/ |
H A D | pkt_sched.h | 486 __u32 gap; /* re-ordering gap (0 for none) */ member in struct:tc_netem_qopt
|