/linux-master/fs/ext4/ |
H A D | orphan.c | 229 struct list_head *prev; local 256 prev = ei->i_orphan.prev; 269 if (prev == &sbi->s_orphan) { 287 &list_entry(prev, struct ext4_inode_info, i_orphan)->vfs_inode;
|
/linux-master/fs/ |
H A D | char_dev.c | 100 struct char_device_struct *cd, *curr, *prev = NULL; local 134 for (curr = chrdevs[i]; curr; prev = curr, curr = curr->next) { 155 if (!prev) { 159 cd->next = prev->next; 160 prev->next = cd;
|
/linux-master/fs/afs/ |
H A D | server.c | 478 struct afs_server *server, *next, *prev; local 492 prev = server->uuid_prev; 493 if (!prev) { 505 rcu_assign_pointer(prev->uuid_next, next); 507 next->uuid_prev = prev;
|
/linux-master/tools/lib/perf/ |
H A D | evlist.c | 120 perf_evlist__next(struct perf_evlist *evlist, struct perf_evsel *prev) argument 124 if (!prev) { 129 next = list_next_entry(prev, node); 389 struct perf_mmap *prev = i ? &map[i - 1] : NULL; local 400 perf_mmap__init(&map[i], prev, overwrite, NULL);
|
/linux-master/drivers/scsi/ |
H A D | scsi.c | 785 struct scsi_device *prev) 787 struct list_head *list = (prev ? &prev->siblings : &shost->__devices); 802 if (prev) 803 scsi_device_put(prev); 784 __scsi_iterate_devices(struct Scsi_Host *shost, struct scsi_device *prev) argument
|
/linux-master/arch/s390/kernel/ |
H A D | perf_pai_crypto.c | 314 u64 prev, new, delta; local 316 prev = local64_read(&event->hw.prev_count); 319 delta = (prev <= new) ? new - prev 320 : (-1ULL - prev) + new + 1; /* overflow */
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dp.c | 357 u8 prev[2]; local 376 memcpy(prev, &stat[4], sizeof(prev)); 381 if (!memcmp(prev, &stat[4], sizeof(prev))) {
|
/linux-master/kernel/ |
H A D | jump_label.c | 740 struct static_key_mod *jlm, **prev; local 759 prev = &key->next; 763 prev = &jlm->next; 771 if (prev == &key->next) 774 *prev = jlm->next;
|
/linux-master/drivers/perf/ |
H A D | dwc_pcie_pmu.c | 342 u64 delta, prev, now = 0; local 345 prev = local64_read(&hwc->prev_count); 352 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); 354 delta = (now - prev) & DWC_PCIE_MAX_PERIOD;
|
/linux-master/fs/reiserfs/ |
H A D | item_ops.c | 264 __u32 prev = INT_MAX; local 274 if (sequence_finished(prev, &num, get_block_num(unp, j))) { 275 print_sequence(prev, num); 276 start_new_sequence(&prev, &num, get_block_num(unp, j)); 279 print_sequence(prev, num);
|
/linux-master/drivers/usb/host/ |
H A D | ehci-q.c | 496 if (stopped && qtd->qtd_list.prev != &qh->qtd_list) { 497 last = list_entry (qtd->qtd_list.prev, 1086 qtd = list_entry (qh->qtd_list.prev, 1260 struct ehci_qh *prev; local 1267 prev = ehci->async; 1268 while (prev->qh_next.qh != qh) 1269 prev = prev->qh_next.qh; 1271 prev->hw->hw_next = qh->hw->hw_next; 1272 prev [all...] |
/linux-master/arch/arc/kernel/ |
H A D | unwind.c | 411 struct unwind_table *table = info->table, *prev; local 413 for (prev = &root_table; prev->link && prev->link != table; 414 prev = prev->link) 417 if (prev->link) { 423 prev->link = table->link; 424 if (!prev->link) 425 last_table = prev; [all...] |
/linux-master/kernel/sched/ |
H A D | psi.c | 916 void psi_task_switch(struct task_struct *prev, struct task_struct *next, argument 920 int cpu = task_cpu(prev); 927 * ancestors with @prev, those will already have @prev's 942 if (prev->pid) { 954 if (prev->in_memstall) 956 if (prev->in_iowait) 965 if (unlikely((prev->flags & PF_WQ_WORKER) && 966 wq_worker_last_func(prev) == psi_avgs_work)) 970 psi_flags_change(prev, clea [all...] |
/linux-master/net/rfkill/ |
H A D | core.c | 316 bool prev, curr; local 331 prev = rfkill->state & RFKILL_BLOCK_SW; 333 if (prev) 368 if (prev != curr) 545 bool ret, prev; local 555 prev = !!(rfkill->hard_block_reasons & reason); 570 if (rfkill->registered && prev != blocked) 594 bool prev, hwblock; local 599 prev = !!(rfkill->state & RFKILL_BLOCK_SW); 608 if (prev ! [all...] |
/linux-master/include/linux/ |
H A D | of.h | 228 extern struct device_node *__of_find_all_nodes(struct device_node *prev); 229 extern struct device_node *of_find_all_nodes(struct device_node *prev); 291 struct device_node *prev); 293 const struct device_node *node, struct device_node *prev); 295 const struct device_node *node, struct device_node *prev); 359 extern struct device_node *of_get_next_cpu_node(struct device_node *prev); 542 const struct device_node *node, struct device_node *prev) 548 const struct device_node *node, struct device_node *prev) 554 const struct device_node *node, struct device_node *prev) 658 static inline struct device_node *of_get_next_cpu_node(struct device_node *prev) argument 541 of_get_next_child( const struct device_node *node, struct device_node *prev) argument 547 of_get_next_available_child( const struct device_node *node, struct device_node *prev) argument 553 of_get_next_reserved_child( const struct device_node *node, struct device_node *prev) argument [all...] |
/linux-master/sound/pci/au88x0/ |
H A D | au88x0_core.c | 261 int temp, lifeboat = 0, prev; local 269 prev = VORTEX_MIXER_CHNBASE + (ch << 2); 270 temp = hwread(vortex->mmio, prev); 272 prev = VORTEX_MIXER_RTBASE + ((temp & 0xf) << 2); 273 temp = hwread(vortex->mmio, prev); 274 //printk(KERN_INFO "vortex: mixAddWTD: while addr=%x, val=%x\n", prev, temp); 282 hwwrite(vortex->mmio, prev, (temp & 0xf) | 0x10); 657 int temp, lifeboat = 0, prev; local 666 prev = VORTEX_SRC_CHNBASE + (ch << 2); /*ebp */ 667 temp = hwread(vortex->mmio, prev); 1632 int temp, prev, lifeboat = 0; local 1679 int temp, lifeboat = 0, prev; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mcg.c | 685 * If GID is found in MGM or MGM is empty, *index = *hash, *prev = -1 688 * if GID is found in AMGM, *index = index in AMGM, *prev = index of 691 * If no AMGM exists for given gid, *index = -1, *prev = index of last 697 int *prev, int *index) 723 *prev = -1; 742 *prev = *index; 1112 int index = -1, prev; local 1126 mailbox, &prev, &index); 1180 err = mlx4_READ_ENTRY(dev, prev, mailbox); 1186 err = mlx4_WRITE_ENTRY(dev, prev, mailbo 694 find_entry(struct mlx4_dev *dev, u8 port, u8 *gid, enum mlx4_protocol prot, struct mlx4_cmd_mailbox *mgm_mailbox, int *prev, int *index) argument 1221 int prev, index; local [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | process.c | 1089 static inline void __switch_to_tm(struct task_struct *prev, argument 1093 if (tm_enabled(prev) || tm_enabled(new)) 1096 if (tm_enabled(prev)) { 1097 prev->thread.load_tm++; 1098 tm_reclaim_task(prev); 1099 if (!MSR_TM_ACTIVE(prev->thread.regs->msr) && prev->thread.load_tm == 0) 1100 prev->thread.regs->msr &= ~MSR_TM; 1152 #define __switch_to_tm(prev, new) 1281 struct task_struct *__switch_to(struct task_struct *prev, 1274 __switch_to(struct task_struct *prev, struct task_struct *new) argument [all...] |
/linux-master/kernel/trace/ |
H A D | ring_buffer.c | 1124 * head->list->prev->next bit 1 bit 0 1130 * Note we can not trust the prev pointer of the head page, because: 1211 struct list_head *list = page->list.prev; 1242 rb_set_list_to_head(head->list.prev); 1269 struct buffer_page *prev, 1276 list = &prev->list; 1292 struct buffer_page *prev, 1295 return rb_head_page_set(cpu_buffer, head, prev, 1301 struct buffer_page *prev, 1304 return rb_head_page_set(cpu_buffer, head, prev, 1267 rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag, int new_flag) argument 1290 rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) argument 1299 rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) argument 1308 rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) argument [all...] |
/linux-master/drivers/md/ |
H A D | raid10.c | 652 geo = &conf->prev; 662 /* Never use conf->prev as this is only called during resync 1831 sector_t chunk_mask = (conf->geo.chunk_mask & conf->prev.chunk_mask); 1853 || conf->prev.near_copies < 1854 conf->prev.raid_disks))) 1905 disks = conf->prev.raid_disks; 1906 ncopies = conf->prev.near_copies; 1935 /* when calling 'enough', both 'prev' and 'geo' must 2969 list_move(conf->bio_end_io_list.prev, &tmp); 2998 r10_bio = list_entry(head->prev, struc [all...] |
/linux-master/fs/nilfs2/ |
H A D | segment.c | 1146 ii->i_dirty.prev, 1168 list_entry(ii->i_dirty.prev, 1274 struct nilfs_segment_buffer *segbuf, *prev; local 1298 prev = NILFS_LAST_SEGBUF(&sci->sc_write_logs); 1299 nilfs_segbuf_map_cont(segbuf, prev); 1300 segbuf->sb_sum.seg_seq = prev->sb_sum.seg_seq; 1301 nextnum = prev->sb_nextnum; 1304 nilfs_segbuf_map(segbuf, prev->sb_nextnum, 0, nilfs); 1334 struct nilfs_segment_buffer *segbuf, *prev; local 1340 prev 1390 struct nilfs_segment_buffer *segbuf, *prev; local [all...] |
/linux-master/drivers/greybus/ |
H A D | operation.c | 137 int prev; local 148 prev = operation->errno; 149 if (prev == -EBADR) 154 WARN_ON(prev != -EBADR); 172 prev = operation->errno; 173 if (prev == -EINPROGRESS) 177 return prev == -EINPROGRESS;
|
/linux-master/drivers/infiniband/core/ |
H A D | rw.c | 128 struct rdma_rw_reg_ctx *prev = NULL; local 150 if (prev) { 152 prev->wr.wr.next = ®->inv_wr; 154 prev->wr.wr.next = ®->reg_wr.wr; 177 prev = reg; 181 if (prev) 182 prev->wr.wr.next = NULL;
|
/linux-master/drivers/dma/xilinx/ |
H A D | zynqmp_dma.c | 300 * @prev: Previous hw descriptor pointer 305 struct zynqmp_dma_desc_ll *prev) 319 if (prev) { 322 ddesc = prev + 1; 323 prev->nxtdscraddr = addr; 826 void *desc = NULL, *prev = NULL; local 851 dma_dst, copy, prev); 852 prev = desc; 302 zynqmp_dma_config_sg_ll_desc(struct zynqmp_dma_chan *chan, struct zynqmp_dma_desc_ll *sdesc, dma_addr_t src, dma_addr_t dst, size_t len, struct zynqmp_dma_desc_ll *prev) argument
|
/linux-master/drivers/comedi/ |
H A D | drivers.c | 1148 struct comedi_driver *prev; local 1156 for (prev = comedi_drivers; prev->next; prev = prev->next) { 1157 if (prev->next == driver) { 1158 prev->next = driver->next;
|