Searched refs:next (Results 201 - 225 of 3417) sorted by relevance

1234567891011>>

/linux-master/arch/arm/mm/
H A Dkasan_init.c46 unsigned long next; local
53 next = addr + PAGE_SIZE;
87 } while (ptep++, addr = next, addr != end);
96 unsigned long next; local
118 next = pmd_addr_end(addr, end);
119 kasan_pte_populate(pmdp, addr, next, early);
120 } while (pmdp++, addr = next, addr != end);
126 unsigned long next; local
149 next = pgd_addr_end(addr, end);
158 kasan_pmd_populate(pudp, addr, next, earl
[all...]
/linux-master/lib/zlib_inflate/
H A Dinftrees.c18 on return points to the next available entry's address. bits is the
40 code *next; /* next available space in table */ local
137 filled is at next and has curr index bits. The code being used is huff
190 next = *table; /* current table to fill in */
221 min = fill; /* save offset to next table */
224 next[(huff >> drop) + fill] = this;
238 /* go to next symbol, update count, len */
252 next += min; /* here min is 1 << curr */
254 /* determine length of next tabl
[all...]
/linux-master/fs/nilfs2/
H A Dsegbuf.h29 * @next: Block number of the next full segment
41 sector_t next; member in struct:nilfs_segsum_info
50 * @sb_nextnum: Index number of the next full segment
87 #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next)
90 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next)
91 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head))
98 (list_entry((head)->next, struct buffer_head, b_assoc_buffers))
100 (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \
102 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next
[all...]
/linux-master/arch/x86/include/asm/
H A Dmmu_context.h76 extern void switch_ldt(struct mm_struct *prev, struct mm_struct *next);
82 static inline void switch_ldt(struct mm_struct *prev, struct mm_struct *next) argument
168 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
171 extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
175 #define activate_mm(prev, next) \
177 paravirt_enter_mmap(next); \
178 switch_mm((prev), (next), NULL); \
/linux-master/drivers/dma-buf/
H A Ddma-fence-unwrap.c42 * dma_fence_unwrap_next - return the next fence from a fence containers
46 * the next fence from them.
115 struct dma_fence *next; local
120 next = fences[i];
121 if (!next)
131 if (!tmp || tmp->context > next->context) {
132 tmp = next;
135 } else if (tmp->context < next->context) {
138 } else if (dma_fence_is_later(tmp, next)) {
/linux-master/kernel/rcu/
H A Dtiny.c33 struct rcu_head **donetail; /* ->next pointer of last "done" CB. */
34 struct rcu_head **curtail; /* ->next pointer of last CB. */
110 struct rcu_head *next, *list; local
130 next = list->next;
131 prefetch(next);
136 list = next;
188 head->next = NULL;
192 rcu_ctrlblk.curtail = &head->next;
/linux-master/kernel/trace/
H A Dpid_list.c22 pid_list->lower_list = chunk->next;
25 chunk->next = NULL;
46 pid_list->upper_list = chunk->next;
49 chunk->next = NULL;
65 chunk->next = pid_list->lower_list;
75 chunk->next = pid_list->upper_list;
262 * trace_pid_list_next - return the next pid in the list
265 * @next: The pointer to place the pid that is set starting from @pid.
267 * Looks for the next consecutive pid that is in @pid_list starting
269 * that pid is placed into @next
273 trace_pid_list_next(struct trace_pid_list *pid_list, unsigned int pid, unsigned int *next) argument
[all...]
/linux-master/drivers/acpi/acpica/
H A Duttrack.c299 if (!element->next) {
303 element = element->next;
383 allocation->next = mem_list->list_head;
390 allocation->next = element->next;
393 if (element->next) {
394 (element->next)->previous = allocation;
397 element->next = allocation;
452 (allocation->previous)->next = allocation->next;
[all...]
/linux-master/fs/omfs/
H A Dfile.c37 u64 next, last; local
45 next = inode->i_ino;
52 bh = omfs_bread(inode->i_sb, next);
61 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next))
69 last = next;
70 next = be64_to_cpu(oe->e_next);
89 if (next == ~0)
92 bh = omfs_bread(inode->i_sb, next);
141 /* trivially grow current extent, if next block is not taken */
221 sector_t next, offse local
[all...]
/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dadf_dev_mgr.c191 struct vf_id_map *next; local
197 next = list_next_entry(map, list);
198 while (next && &next->list != &vfs_table) {
199 next->fake_id++;
200 next = list_next_entry(next, list);
257 struct vf_id_map *map, *next; local
266 next = list_next_entry(map, list);
267 while (next
[all...]
/linux-master/tools/testing/selftests/kvm/lib/
H A Dsparsebit.c224 * If current node has a right child, next node is the left-most
235 * That parent is then the next node.
252 * If current node has a left child, next node is the right-most
263 * That parent is then the next node.
559 * reductions are possible with the new previous and next nodes. Note,
605 struct node *prev, *next, *tmp; local
614 * for the next pass through the reduction loop,
617 * by first remembering the location of the next
620 * no other nodes between prev and next.
623 * both prev and next bot
1373 struct node *nodep, *next; local
1455 struct node *nodep, *next; local
1916 sparsebit_idx_t next; local
[all...]
/linux-master/arch/arm64/mm/
H A Dkasan_init.c118 unsigned long next; local
127 next = addr + PAGE_SIZE;
129 } while (ptep++, addr = next, addr != end && pte_none(__ptep_get(ptep)));
135 unsigned long next; local
139 next = pmd_addr_end(addr, end);
140 kasan_pte_populate(pmdp, addr, next, node, early);
141 } while (pmdp++, addr = next, addr != end && pmd_none(READ_ONCE(*pmdp)));
147 unsigned long next; local
151 next = pud_addr_end(addr, end);
152 kasan_pmd_populate(pudp, addr, next, nod
159 unsigned long next; local
171 unsigned long next; local
[all...]
/linux-master/block/
H A Dblk-merge.c53 struct request *prev_rq, struct bio *prev, struct bio *next)
77 * - if 'pb' ends unaligned, the next bio must include
82 bio_get_first_bvec(next, &nb);
132 * If the next starting sector would be misaligned, stop the discard at
670 struct request *next)
676 if (blk_rq_sectors(req) + bio_sectors(next->bio) >
680 req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
688 struct request *next)
692 if (req_gap_back_merge(req, next->bio))
698 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) >
52 bio_will_gap(struct request_queue *q, struct request *prev_rq, struct bio *prev, struct bio *next) argument
669 req_attempt_discard_merge(struct request_queue *q, struct request *req, struct request *next) argument
687 ll_merge_requests_fn(struct request_queue *q, struct request *req, struct request *next) argument
786 blk_try_req_merge(struct request *req, struct request *next) argument
801 attempt_merge(struct request_queue *q, struct request *req, struct request *next) argument
890 struct request *next = elv_latter_request(q, rq); local
914 blk_attempt_req_merge(struct request_queue *q, struct request *rq, struct request *next) argument
[all...]
/linux-master/fs/
H A Dchar_dev.c35 struct char_device_struct *next; member in struct:char_device_struct
56 for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) {
77 for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next)
134 for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) {
156 cd->next = curr;
159 cd->next = prev->next;
160 prev->next = cd;
178 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next)
185 *cp = cd->next;
204 dev_t n, next; local
314 dev_t n, next; local
[all...]
H A Dfilesystems.c52 for (p = &file_systems; *p; p = &(*p)->next)
82 if (fs->next)
116 *tmp = fs->next;
117 fs->next = NULL;
122 tmp = &(*tmp)->next;
145 for (tmp=file_systems, index=0 ; tmp ; tmp=tmp->next, index++) {
162 for (tmp = file_systems; tmp; tmp = tmp->next, index--)
182 for (tmp = file_systems, index = 0 ; tmp ; tmp = tmp->next, index++)
219 for (p = file_systems; p; p = p->next) {
247 tmp = tmp->next;
[all...]
/linux-master/arch/s390/mm/
H A Dpageattr.c178 unsigned long next; local
187 next = pmd_addr_end(addr, end);
191 need_split |= !!(addr + PMD_SIZE > next);
200 rc = walk_pte_level(pmdp, addr, next, flags);
205 addr = next;
267 unsigned long next; local
276 next = pud_addr_end(addr, end);
280 need_split |= !!(addr + PUD_SIZE > next);
289 rc = walk_pmd_level(pudp, addr, next, flags);
292 addr = next;
301 unsigned long next; local
323 unsigned long next; local
[all...]
/linux-master/drivers/tty/
H A Dtty_buffer.c106 p->next = NULL;
123 struct tty_buffer *p, *next; local
129 buf->head = p->next;
135 llist_for_each_entry_safe(p, next, llist, free)
225 struct tty_buffer *next; local
233 while ((next = smp_load_acquire(&buf->head->next)) != NULL) {
235 buf->head = next;
287 * is advanced to the next buffer.
289 smp_store_release(&b->next,
402 struct tty_buffer *next; local
471 struct tty_buffer *next; local
[all...]
/linux-master/drivers/gpu/drm/radeon/
H A Dmkregtable.c37 * sometimes we already know the next/prev entries and we can
43 struct list_head *next, *prev; member in struct:list_head
49 list->next = list;
57 * the prev/next entries already!
61 struct list_head *prev, struct list_head *next)
63 next->prev = new;
64 new->next = next;
66 prev->next = new;
70 struct list_head *prev, struct list_head *next);
60 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) argument
[all...]
/linux-master/drivers/pcmcia/
H A Drsrc_nonstatic.c50 struct resource_map *next; member in struct:resource_map
112 for (p = map; ; p = p->next) {
117 if ((p->next == map) || (p->next->base > base+num-1))
126 q->next = p->next; p->next = q;
137 q = p->next;
144 p->next = q->next;
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dsubpage_prot.c20 * reinitialized for the next user.
97 unsigned long next, limit; local
108 for (; addr < limit; addr = next) {
109 next = pmd_addr_end(addr, limit);
124 if (addr + (nw << PAGE_SHIFT) > next)
125 nw = (next - addr) >> PAGE_SHIFT;
192 unsigned long next, limit; local
233 for (limit = addr + len; addr < limit; addr = next) {
234 next = pmd_addr_end(addr, limit);
263 if (addr + (nw << PAGE_SHIFT) > next)
[all...]
/linux-master/arch/powerpc/include/asm/
H A Dmmu_context.h56 struct mm_struct *next);
58 struct mm_struct *next,
62 return radix__switch_mmu_context(prev, next);
63 return switch_slb(tsk, next);
100 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
151 * flush of the full mm, as the next TLBI may be local and the
226 extern void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next,
229 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
235 switch_mm_irqs_off(prev, next, tsk);
245 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next)
57 switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) argument
[all...]
/linux-master/arch/riscv/kernel/
H A Dvector.c214 int cur, next; local
219 next = riscv_v_ctrl_get_next(tsk);
220 if (!next) {
226 cur = next;
228 /* Clear next mask if inherit-bit is not set */
231 next = PR_RISCV_V_VSTATE_CTRL_DEFAULT;
233 riscv_v_ctrl_set(tsk, cur, next, inherit);
247 int cur, next; local
272 next = VSTATE_CTRL_GET_NEXT(arg);
274 switch (next) {
[all...]
/linux-master/arch/riscv/mm/
H A Dcontext.c91 /* Queue a TLB invalidation for each CPU on next context-switch */
166 * of active_context as used and next context switch we will
209 struct mm_struct *next, unsigned int cpu)
219 cpumask_set_cpu(cpu, mm_cpumask(next));
221 set_mm_asid(next, cpu);
224 set_mm_noasid(next);
280 struct mm_struct *next, unsigned int cpu)
318 void switch_mm(struct mm_struct *prev, struct mm_struct *next, argument
323 if (unlikely(prev == next))
326 membarrier_arch_switch_mm(prev, next, tas
208 set_mm(struct mm_struct *prev, struct mm_struct *next, unsigned int cpu) argument
279 set_mm(struct mm_struct *prev, struct mm_struct *next, unsigned int cpu) argument
[all...]
/linux-master/net/ax25/
H A Dax25_dev.c37 for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next)
88 ax25_dev->next = ax25_dev_list;
114 for (s = ax25_dev_list; s != NULL; s = s->next)
119 ax25_dev_list = s->next;
123 while (s != NULL && s->next != NULL) {
124 if (s->next == ax25_dev) {
125 s->next = ax25_dev->next;
129 s = s->next;
210 ax25_dev = ax25_dev->next;
[all...]
/linux-master/net/unix/
H A Dgarbage.c163 struct sk_buff *next; local
166 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) {
203 struct sk_buff *next; local
211 skb_queue_walk_safe(&x->sk_receive_queue, skb, next) {
223 u = list_entry(embryos.next, struct unix_sock, link);
257 struct unix_sock *u, *next; local
288 list_for_each_entry_safe(u, next, &gc_inflight_list, link) {
322 while (cursor.next != &gc_candidates) {
323 u = list_entry(cursor.next, struct unix_sock, link);
356 u = list_entry(not_cycle_list.next, struc
[all...]

Completed in 2921 milliseconds

1234567891011>>