/linux-master/arch/powerpc/platforms/powernv/ |
H A D | opal-memory-errors.c | 76 static void mem_error_handler(struct work_struct *work) argument
|
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | os-area.c | 669 static void os_area_queue_work_handler(struct work_struct *work) argument
|
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | Makefile | 5 of_helpers.o rtas-work-area.o papr-sysparm.o \
|
H A D | dlpar.c | 25 #include <asm/rtas-work-area.h> 30 struct work_struct work; member in struct:pseries_hp_work 374 static void pseries_hp_work_fn(struct work_struct *work) argument 377 container_of(work, struct pseries_hp_work, work); 382 kfree(work); 387 struct pseries_hp_work *work; local 394 work = kmalloc(sizeof(struct pseries_hp_work), GFP_ATOMIC); 395 if (work) { 396 INIT_WORK((struct work_struct *)work, pseries_hp_work_f [all...] |
H A D | lpar.c | 144 struct delayed_work work; member in struct:dtl_worker 360 static void process_dtl_buffer(struct work_struct *work) argument 367 struct dtl_worker *d = container_of(work, struct dtl_worker, work.work); 404 schedule_delayed_work_on(d->cpu, to_delayed_work(work), 413 INIT_DELAYED_WORK(&d->work, process_dtl_buffer); 423 schedule_delayed_work_on(cpu, &d->work, HZ / vcpudispatch_stats_freq); 431 cancel_delayed_work_sync(&d->work);
|
H A D | papr-sysparm.c | 17 #include <asm/rtas-work-area.h> 84 * work area, the caller-supplied buffer is copied unmodified into the 85 * work area before calling ibm,get-system-parameter.
|
H A D | papr-vpd.c | 20 #include <asm/rtas-work-area.h> 35 * @work_area: In: Caller-provided work area buffer for results. 49 * rtas_ibm_get_vpd() - Call ibm,get-vpd to fill a work area buffer. 53 * into the supplied work area. Handles RTAS retry statuses. Maps RTAS 106 "possible write beyond end of work area")) 243 * RTAS to ensure it's in the RMA and avoid a separate work 249 * We could allocate the work area before acquiring the 251 * exhaust the limited work area pool for no benefit. So 252 * allocate the work area under the lock. 349 * Context: May sleep. Holds a mutex and an RTAS work are [all...] |
H A D | rtas-work-area.c | 3 #define pr_fmt(fmt) "rtas-work-area: " fmt 17 #include <asm/rtas-work-area.h> 30 * The smallest known work area size is for ibm,get-vpd's 53 * A single work area buffer and descriptor to serve requests early in 127 * Initialization of the work area allocator happens in two parts. To 151 * All RTAS functions that consume work areas are OK with 186 * rtas_work_area_reserve_arena() - Reserve memory suitable for RTAS work areas.
|
H A D | vio.c | 57 /* How many ms to delay queued balance work */ 78 * @balance_q: work queue for balancing system entitlement 368 * @work: work queue structure for this operation 386 static void vio_cmo_balance(struct work_struct *work) argument 395 cmo = container_of(work, struct vio_cmo, balance_q.work); 478 cancel_delayed_work(to_delayed_work(work));
|
/linux-master/arch/powerpc/sysdev/ |
H A D | fsl_rmu.c | 479 static void fsl_pw_dpc(struct work_struct *work) argument 481 struct fsl_rio_pw *pw = container_of(work, struct fsl_rio_pw, pw_work);
|
H A D | pmi.c | 38 struct work_struct work; member in struct:pmi_data 84 schedule_work(&data->work); 100 static void pmi_notify_handlers(struct work_struct *work) argument 144 INIT_WORK(&data->work, pmi_notify_handlers);
|
/linux-master/arch/s390/appldata/ |
H A D | appldata_base.c | 81 static void appldata_work_fn(struct work_struct *work); 92 /*************************** timer, work, DIAG *******************************/ 96 * schedule work and reschedule timer 108 static void appldata_work_fn(struct work_struct *work) argument 156 /************************ timer, work, DIAG <END> ****************************/ 307 // protect work queue callback
|
/linux-master/arch/s390/include/asm/ |
H A D | kvm_host.h | 1018 struct kvm_async_pf *work); 1021 struct kvm_async_pf *work); 1024 struct kvm_async_pf *work);
|
/linux-master/arch/s390/kernel/ |
H A D | time.c | 442 static void stp_work_fn(struct work_struct *work); 585 /* Didn't work. Clear per-cpu in sync bit again. */ 669 * STP work. Check for the STP state and take over the clock 672 static void stp_work_fn(struct work_struct *work) argument
|
H A D | topology.c | 48 static void topology_work_fn(struct work_struct *work); 336 static void topology_work_fn(struct work_struct *work) argument
|
/linux-master/arch/s390/kvm/ |
H A D | kvm-s390.c | 3963 * cpu_has_vx(), (load|store)_fpu_regs() will work with vrs format. 4611 struct kvm_async_pf *work) 4613 trace_kvm_s390_pfault_init(vcpu, work->arch.pfault_token); 4614 __kvm_inject_pfault_token(vcpu, true, work->arch.pfault_token); 4620 struct kvm_async_pf *work) 4622 trace_kvm_s390_pfault_done(vcpu, work->arch.pfault_token); 4623 __kvm_inject_pfault_token(vcpu, false, work->arch.pfault_token); 4627 struct kvm_async_pf *work) 4610 kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) argument 4619 kvm_arch_async_page_present(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) argument 4626 kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) argument
|
/linux-master/arch/sh/drivers/ |
H A D | push-switch.c | 30 schedule_work(&psw->work); 33 static void switch_work_handler(struct work_struct *work) argument 35 struct push_switch *psw = container_of(work, struct push_switch, work); 77 INIT_WORK(&psw->work, switch_work_handler); 105 flush_work(&psw->work);
|
/linux-master/arch/sparc/kernel/ |
H A D | leon_smp.c | 138 prom_printf("######## !!!! The irqmp-ctrl must have broadcast enabled, smp wont work !!!!! ####### nr cpus: %d\n", 275 struct leon_ipi_work *work; local 298 work = &per_cpu(leon_ipi_work, cpu); 299 work->single = work->msk = work->resched = 0; 312 struct leon_ipi_work *work = &per_cpu(leon_ipi_work, cpu); local 314 /* Mark work */ 315 work->single = 1; 323 struct leon_ipi_work *work local 334 struct leon_ipi_work *work = &per_cpu(leon_ipi_work, cpu); local 345 struct leon_ipi_work *work = this_cpu_ptr(&leon_ipi_work); local [all...] |
H A D | sun4d_smp.c | 196 struct sun4d_ipi_work *work; local 201 work = &per_cpu(sun4d_ipi_work, cpu); 202 work->single = work->msk = work->resched = 0; 208 struct sun4d_ipi_work *work = this_cpu_ptr(&sun4d_ipi_work); local 210 if (work->single) { 211 work->single = 0; 214 if (work->msk) { 215 work 239 struct sun4d_ipi_work *work = &per_cpu(sun4d_ipi_work, cpu); local 250 struct sun4d_ipi_work *work = &per_cpu(sun4d_ipi_work, cpu); local 261 struct sun4d_ipi_work *work = &per_cpu(sun4d_ipi_work, cpu); local [all...] |
/linux-master/arch/um/drivers/ |
H A D | chan_kern.c | 131 static void line_timer_cb(struct work_struct *work) argument 133 struct line *line = container_of(work, struct line, task.work);
|
H A D | line.c | 596 struct work_struct work; member in struct:winch 599 static void __free_winch(struct work_struct *work) argument 601 struct winch *winch = container_of(work, struct winch, work); 617 __free_winch(&winch->work); 644 INIT_WORK(&winch->work, __free_winch); 645 schedule_work(&winch->work);
|
H A D | net_kern.c | 108 static void uml_dev_close(struct work_struct *work) argument 111 container_of(work, struct uml_net_private, work); 136 * enqueued work item. */ 137 schedule_work(&lp->work); 410 INIT_WORK(&lp->work, uml_dev_close);
|
H A D | vector_kern.c | 1196 static void vector_reset_tx(struct work_struct *work) argument 1199 container_of(work, struct vector_private, reset_tx);
|
/linux-master/arch/um/include/shared/ |
H A D | net_kern.h | 29 struct work_struct work; member in struct:uml_net_private
|
/linux-master/arch/x86/boot/ |
H A D | header.S | 545 # which happened to work by accident for the old code. Recalculate the stack
|