/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_tc.c | 114 * - change duration of last operation 212 /* last operation of list is reserved to insert operation */ 288 u64 last; local 291 last = tsnep_gcl_start_before(curr, gcl->start_time); 292 if ((last + curr->cycle_time) == gcl->start_time) 293 change = tsnep_cut_gcl(curr, last, 294 gcl->start_time - last); 295 else if (((gcl->start_time - last) <= 297 ((gcl->start_time - last) <= TSNEP_GCL_MIN_INTERVAL)) 298 change = tsnep_extend_gcl(curr, last, [all...] |
/linux-master/fs/bcachefs/ |
H A D | movinggc.c | 32 struct move_bucket_in_flight *last; member in struct:buckets_in_flight 64 list->last->next = new; 66 list->last = new; 120 list->last = NULL; 313 u64 last, wait; local 351 last = atomic64_read(&clock->now); 355 c->copygc_wait_at = last; 356 c->copygc_wait = last + wait; 358 trace_and_count(c, copygc_wait, c, wait, last + wait); 359 bch2_kthread_io_clock_wait(clock, last [all...] |
/linux-master/arch/mips/jazz/ |
H A D | jazzdma.c | 97 int first, last, pages, frame, i; local 128 last = first + 1; 129 while (pgtbl[last].owner == VDMA_PAGE_EMPTY 130 && last - first < pages) 131 last++; 133 if (last - first == pages) 135 first = last + 1; 144 for (i = first; i < last; i++) { 161 for (i = first; i < last; i++) 164 for (i = first; i < last; [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | switch_to.h | 109 #define switch_to(prev, next, last) \ 139 (last) = resume(prev, next, task_thread_info(next)); \
|
/linux-master/arch/loongarch/lib/ |
H A D | dump_tlb.c | 28 static void dump_tlb(int first, int last) argument 44 for (i = first; i <= last; i++) {
|
/linux-master/arch/riscv/include/asm/ |
H A D | switch_to.h | 75 #define switch_to(prev, next, last) \ 83 ((last) = __switch_to(__prev, __next)); \
|
/linux-master/arch/powerpc/include/asm/ |
H A D | switch_to.h | 17 #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next)))
|
/linux-master/fs/adfs/ |
H A D | dir_f.c | 76 int last = 5 - 26; local 80 * Accumulate each word up to the last whole 81 * word of the last directory entry. This 85 last += 26; 90 } while (i < (last & ~3)); 91 } while (dir_u8(last) != 0); 94 * Accumulate the last few bytes. These 97 if (i != last) { 99 end.ptr8 = ptr.ptr8 + last - i;
|
/linux-master/lib/zlib_inflate/ |
H A D | inffast.c | 71 const unsigned char *last; /* while in < last, enough input available */ local 98 last = in + (strm->avail_in - 5); 308 } while (in < last && out < end); 319 strm->avail_in = (unsigned)(in < last ? 5 + (last - in) : 5 - (in - last));
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | radix-sort.c | 35 /* The index (key byte) of the last non-empty bin */ 36 u16 last; member in struct:histogram 48 /* Pointer to the last key to sort. */ 82 /* Insert the key into the last slot that was cleared, sorting it. */ 119 * offset, keeping track of the number of non-empty bins, and the index of the first and last 132 bins->last = 0; 145 if (bin > bins->last) 146 bins->last = bin; 293 * Don't bother processing the last pile: when piles 0..N-1 are all in place, then 296 end = task.last_key - bins->size[bins->last]; [all...] |
/linux-master/drivers/pwm/ |
H A D | pwm-stmpe.c | 172 u8 value, last = stmpe_pwm->last_duty; local 190 if (value == last) { 204 if (last < value) 206 incdec = RAMPUP | (value - last); 209 incdec = RAMPDOWN | (last - value); 220 pwm->hwpwm, value, last, program[0], program[1], local
|
/linux-master/include/linux/ |
H A D | rculist.h | 210 * @prev: points to the last element of the existing list 232 struct list_head *last = list->prev; local 235 * "first" and "last" tracking list, so initialize it. RCU readers 251 ASSERT_EXCLUSIVE_ACCESS(*last); 261 last->next = next; 264 next->prev = last; 618 struct hlist_node *i, *last = NULL; local 622 last = i; 624 if (last) { 625 n->next = last [all...] |
/linux-master/lib/ |
H A D | maple_tree.c | 45 * a slot, but the last offset has an implied pivot from the node above (or 391 * a reuse of the last bit in the node type. This is possible by using bit 1 to 529 * Okay to use MAPLE_PARENT_16B_SLOT_MASK as the last bit will be lost 1405 * Return: The zero indexed last slot with data (may be null). 1435 * Return: The zero indexed last slot with data (may be null). 2098 b_node->pivot[b_end] = mas->last; 2101 if (mas->last >= mas->max) 2106 if (piv > mas->last) { 2116 b_node->gap[b_end] = piv - mas->last + 1; 2163 MA_STATE(parent, mas->tree, mas->index, mas->last); 2756 struct maple_enode *next, *last; local 3542 unsigned long last = wr_mas->mas->last; local 4986 struct maple_enode *last = NULL; local 5104 struct maple_enode *last = mas->node; local 6359 mtree_store_range(struct maple_tree *mt, unsigned long index, unsigned long last, void *entry, gfp_t gfp) argument 6414 mtree_insert_range(struct maple_tree *mt, unsigned long first, unsigned long last, void *entry, gfp_t gfp) argument 7150 unsigned long last = max; local 7213 unsigned long last = max; local 7531 void *entry, *last = (void *)1; local [all...] |
/linux-master/net/ethtool/ |
H A D | cabletest.c | 237 cfg->last = MAX_CABLE_LENGTH_CM; 255 cfg->last = nla_get_u32(tb[ETHTOOL_A_CABLE_TEST_TDR_CFG_LAST]); 278 if (cfg->last > MAX_CABLE_LENGTH_CM) { 281 "invalid last parameter"); 285 if (cfg->first > cfg->last) { 286 NL_SET_ERR_MSG(info->extack, "invalid first/last parameter"); 297 if (cfg->step > (cfg->last - cfg->first)) { 406 int ethnl_cable_test_step(struct phy_device *phydev, u32 first, u32 last, argument 420 if (nla_put_u32(phydev->skb, ETHTOOL_A_CABLE_STEP_LAST_DISTANCE, last))
|
/linux-master/drivers/i2c/busses/ |
H A D | i2c-digicolor.c | 55 int last; member in struct:dc_i2c 104 bool last = (i2c->msgbuf_ptr + 1 == i2c->msg->len); local 106 dc_i2c_cmd(i2c, last ? II_CMD_GET_NOACK : II_CMD_GET_ACK); 112 if (i2c->last) 214 int last) 222 i2c->last = last; 213 dc_i2c_xfer_msg(struct dc_i2c *i2c, struct i2c_msg *msg, int first, int last) argument
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | intel_memory_region.c | 92 static resource_size_t random_page(resource_size_t last) argument 95 return get_random_u32_below(last >> PAGE_SHIFT) << PAGE_SHIFT; 102 resource_size_t last, page; local 108 last = resource_size(&mem->io) - PAGE_SIZE; 116 * the first and last of the specified region to confirm the backing 122 for (page = 0; page <= last; page += PAGE_SIZE) { 132 err = iopagetest(mem, last, caller); 136 err = iopagetest(mem, random_page(last), caller);
|
/linux-master/tools/testing/selftests/net/tcp_ao/ |
H A D | rst.c | 261 unsigned int last = ARRAY_SIZE(sk) - 1; local 273 for (i = 0; i < last; i++) { 281 err = test_wait_fds(sk, last, is_writable, last, TEST_TIMEOUT_SEC); 286 err = _test_connect_socket(sk[last], this_ip_dest, port, -1); 304 err = test_wait_fds(sk, last, NULL, last, TEST_TIMEOUT_SEC);
|
/linux-master/arch/arm64/kernel/ |
H A D | fpsimd.c | 60 * (a) for each task, we need to remember which CPU was the last one to have 430 * current context is the context last bound to the CPU stored in 431 * last, if KVM is involved this may be the guest VM context rather 433 * that we must always reference the state storage via last rather 435 * ensured that the type of registers to save is set in last->to_save. 439 struct cpu_fp_state const *last = local 453 *(last->fpmr) = read_sysreg_s(SYS_FPMR); 460 if ((last->to_save == FP_STATE_CURRENT && test_thread_flag(TIF_SVE) && 462 last->to_save == FP_STATE_SVE) { 465 vl = last 1512 struct cpu_fp_state *last = this_cpu_ptr(&fpsimd_last_state); local 1707 struct cpu_fp_state *last = this_cpu_ptr(&fpsimd_last_state); local 1742 struct cpu_fp_state *last = this_cpu_ptr(&fpsimd_last_state); local [all...] |
/linux-master/arch/x86/events/intel/ |
H A D | pt.c | 571 * @last: index of the last initialized entry in this table 578 int last; member in struct:topa 615 /* make -1 stand for the last table entry */ 618 ? &topa_to_page(t)->table[(t)->last] \ 668 tp->topa.last = 0; 698 * "last" table and adjust the last table pointer to @topa. 702 struct topa *last = buf->last; local [all...] |
/linux-master/net/core/ |
H A D | datagram.c | 85 * Wait for the last received packet to be different from skb 170 struct sk_buff **last) 181 *last = queue->prev; 214 * @last: set to last peeked message to inform the wait function 233 * * (BTW I've just killed the last cli() in IP/IPv6/core/netlink/packet 244 struct sk_buff **last) 265 last); 276 } while (READ_ONCE(queue->prev) != *last); 290 struct sk_buff *skb, *last; local 166 __skb_try_recv_from_queue(struct sock *sk, struct sk_buff_head *queue, unsigned int flags, int *off, int *err, struct sk_buff **last) argument 241 __skb_try_recv_datagram(struct sock *sk, struct sk_buff_head *queue, unsigned int flags, int *off, int *err, struct sk_buff **last) argument 669 skb_frag_t *last = &skb_shinfo(skb)->frags[frag - 1]; local [all...] |
/linux-master/net/netfilter/ipvs/ |
H A D | ip_vs_wrr.c | 41 * last pass: cw = di 48 * so that last cw is 1 to catch such dests with weight below di: 53 * last pass: cw = 1 163 struct ip_vs_dest *dest, *last, *stop = NULL; local 174 last = dest; 175 /* Stop only after all dests were checked for weight >= 1 (last pass) */ 189 /* Stop if we tried last pass from first dest: 192 * 2. last was head: the first and only traversal 196 &last->n_list == &svc->destinations) 202 &last [all...] |
/linux-master/fs/f2fs/ |
H A D | xattr.c | 632 struct f2fs_xattr_entry *here, *last; local 696 last = here; 697 while (!IS_XATTR_LAST_ENTRY(last)) { 698 if ((void *)(last) + sizeof(__u32) > last_base_addr || 699 (void *)XATTR_NEXT_ENTRY(last) > last_base_addr) { 700 f2fs_err(F2FS_I_SB(inode), "inode (%lu) has invalid last xattr entry, entry_size: %zu", 701 inode->i_ino, ENTRY_SIZE(last)); 708 last = XATTR_NEXT_ENTRY(last); 720 free = MIN_OFFSET(inode) - ((char *)last [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | errors.c | 73 int last; member in struct:error_block 90 .last = UDS_ERROR_CODE_LAST, 109 if ((errnum >= block->base) && (errnum < block->last)) { 112 } else if ((errnum >= block->last) && (errnum < block->max)) { 279 .last = first_error + (info_size / sizeof(struct error_info)),
|
/linux-master/arch/alpha/kernel/ |
H A D | core_irongate.c | 317 unsigned long baddr, last; 357 last = addr + size - 1; 358 size = PAGE_ALIGN(last) - addr; 367 for(baddr = addr; baddr <= last; baddr += PAGE_SIZE) 383 baddr <= last; 316 unsigned long baddr, last; local
|
/linux-master/kernel/irq/ |
H A D | msi.c | 26 * @last: Last (hardware) slot index to operate on 33 unsigned int last; member in struct:msi_ctrl 173 if (WARN_ON_ONCE(ctrl->first > ctrl->last || 175 ctrl->last >= hwsize)) 192 xa_for_each_range(xa, idx, desc, ctrl->first, ctrl->last) { 207 * @last: Last index to be freed (inclusive) 210 unsigned int first, unsigned int last) 215 .last = last, 239 for (idx = ctrl->first; idx <= ctrl->last; id 209 msi_domain_free_msi_descs_range(struct device *dev, unsigned int domid, unsigned int first, unsigned int last) argument 1404 msi_domain_alloc_irqs_range_locked(struct device *dev, unsigned int domid, unsigned int first, unsigned int last) argument 1427 msi_domain_alloc_irqs_range(struct device *dev, unsigned int domid, unsigned int first, unsigned int last) argument 1656 msi_domain_free_irqs_range_locked(struct device *dev, unsigned int domid, unsigned int first, unsigned int last) argument 1676 msi_domain_free_irqs_range(struct device *dev, unsigned int domid, unsigned int first, unsigned int last) argument [all...] |