/linux-master/block/ |
H A D | blk-zoned.c | 569 static void blk_zone_wplug_bio_work(struct work_struct *work); 698 * plugs work list, increase its reference count. This reference will 720 * We are racing with the error handling work which drops the reference 1116 * plug bio submit work. 1192 * If we had an error, schedule error recovery. The recovery work 1290 static void blk_zone_wplug_bio_work(struct work_struct *work) argument 1293 container_of(work, struct blk_zone_wplug, bio_work); 1427 static void disk_zone_wplugs_work(struct work_struct *work) argument 1430 container_of(work, struct gendisk, zone_wplugs_work);
|
H A D | disk-events.c | 84 * the event work before further blockers are allowed to finish. 307 static void disk_events_workfn(struct work_struct *work) argument 309 struct delayed_work *dwork = to_delayed_work(work);
|
/linux-master/crypto/ |
H A D | cryptd.c | 38 struct work_struct work; member in struct:cryptd_cpu_queue 98 static void cryptd_queue_worker(struct work_struct *work); 112 INIT_WORK(&cpu_queue->work, cryptd_queue_worker); 146 queue_work_on(smp_processor_id(), cryptd_wq, &cpu_queue->work); 159 /* Called in workqueue context, do one real cryption work (via 160 * req->complete) and reschedule itself if there are more work to 162 static void cryptd_queue_worker(struct work_struct *work) argument 167 cpu_queue = container_of(work, struct cryptd_cpu_queue, work); 184 queue_work(cryptd_wq, &cpu_queue->work); [all...] |
H A D | crypto_engine.c | 222 static void crypto_pump_work(struct kthread_work *work) argument 225 container_of(work, struct crypto_engine, pump_requests);
|
/linux-master/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 52 static void job_wq_completion(struct work_struct *work); 521 * in such cases just continue with the cs to cancel it's TDR work. 816 static void cs_timedout(struct work_struct *work) argument 818 struct hl_cs *cs = container_of(work, struct hl_cs, work_tdr.work); 1166 static void job_wq_completion(struct work_struct *work) argument 1168 struct hl_cs_job *job = container_of(work, struct hl_cs_job, 1177 static void cs_completion(struct work_struct *work) argument 1179 struct hl_cs *cs = container_of(work, struct hl_cs, finish_work);
|
H A D | decoder.c | 46 static void dec_abnrm_intr_work(struct work_struct *work) argument 48 struct hl_dec *dec = container_of(work, struct hl_dec, abnrm_intr_work);
|
H A D | device.c | 776 static void device_hard_reset_pending(struct work_struct *work) argument 779 container_of(work, struct hl_device_reset_work, reset_work.work); 809 static void device_release_watchdog_func(struct work_struct *work) argument 812 container_of(work, struct hl_device_reset_work, reset_work.work); 1065 static void hl_device_heartbeat(struct work_struct *work) argument 1067 struct hl_device *hdev = container_of(work, struct hl_device, 1068 work_heartbeat.work); 1099 * If control reached here, then at least one heartbeat work ha [all...] |
H A D | irq.c | 13 * struct hl_eqe_work - This structure is used to schedule work of EQ 58 static void irq_handle_eqe(struct work_struct *work) argument 60 struct hl_eqe_work *eqe_work = container_of(work, struct hl_eqe_work, 70 * job_finish - queue job finish work 92 * cs_finish - queue all cs jobs finish work 201 * @work: workqueue object pointer 203 static void hl_ts_free_objects(struct work_struct *work) argument 206 container_of(work, struct timestamp_reg_work_obj, free_obj);
|
/linux-master/drivers/accel/habanalabs/common/mmu/ |
H A D | mmu.c | 668 static void hl_mmu_prefetch_work_function(struct work_struct *work) argument 670 struct hl_prefetch_work *pfw = container_of(work, struct hl_prefetch_work, prefetch_work); 709 * at the end of the work function)
|
/linux-master/drivers/accel/habanalabs/goya/ |
H A D | goya.c | 832 static void goya_set_freq_to_low_job(struct work_struct *work) argument 834 struct goya_work_freq *goya_work = container_of(work, 836 work_freq.work);
|
/linux-master/drivers/accel/ivpu/ |
H A D | ivpu_pm.c | 112 static void ivpu_pm_recovery_work(struct work_struct *work) argument 114 struct ivpu_pm_info *pm = container_of(work, struct ivpu_pm_info, recovery_work); 169 static void ivpu_job_timeout_work(struct work_struct *work) argument 171 struct ivpu_pm_info *pm = container_of(work, struct ivpu_pm_info, job_timeout_work.work);
|
/linux-master/drivers/accel/qaic/ |
H A D | qaic.h | 107 /* Polling work item to simulate interrupts */ 306 void irq_polling_work(struct work_struct *work);
|
H A D | qaic_control.c | 235 struct work_struct work; member in struct:resp_work 1397 static void resp_worker(struct work_struct *work) argument 1399 struct resp_work *resp = container_of(work, struct resp_work, work); 1463 INIT_WORK(&resp->work, resp_worker); 1466 queue_work(qdev->cntl_wq, &resp->work);
|
H A D | qaic_data.c | 1482 void irq_polling_work(struct work_struct *work) argument 1484 struct dma_bridge_chan *dbc = container_of(work, struct dma_bridge_chan, poll_work);
|
H A D | qaic_debugfs.c | 32 /* Work struct to schedule work coming on QAIC_LOGGING channel */ 33 struct work_struct work; member in struct:bootlog_msg 215 static void bootlog_log(struct work_struct *work) argument 217 struct bootlog_msg *msg = container_of(work, struct bootlog_msg, work); 262 INIT_WORK(&msg->work, bootlog_log); 311 queue_work(qdev->bootlog_wq, &msg->work);
|
H A D | qaic_timesync.c | 93 struct work_struct work; member in struct:qts_resp 248 static void qaic_boot_timesync_worker(struct work_struct *work) argument 250 struct qts_resp *resp = container_of(work, struct qts_resp, work); 303 INIT_WORK(&resp->work, qaic_boot_timesync_worker); 360 queue_work(resp->qdev->qts_wq, &resp->work);
|
H A D | sahara.c | 89 struct work_struct work; member in struct:sahara_context 323 static void sahara_processing(struct work_struct *work) argument 325 struct sahara_context *context = container_of(work, struct sahara_context, work); 385 INIT_WORK(&context->work, sahara_processing); 408 cancel_work_sync(&context->work); 422 schedule_work(&context->work);
|
/linux-master/drivers/accessibility/speakup/ |
H A D | selection.c | 20 struct work_struct work; member in struct:speakup_selection_work 25 static void __speakup_set_selection(struct work_struct *work) argument 28 container_of(work, struct speakup_selection_work, work); 58 .work = __WORK_INITIALIZER(speakup_sel_work.work, 65 * cancelling selection work. getting kref first establishes the 87 schedule_work_on(WORK_CPU_UNBOUND, &speakup_sel_work.work); 96 cancel_work_sync(&speakup_sel_work.work); 97 /* setting to null so that if work fail 106 __speakup_paste_selection(struct work_struct *work) argument [all...] |
/linux-master/drivers/acpi/ |
H A D | acpi_video.c | 219 static void acpi_video_switch_brightness(struct work_struct *work); 478 * key event reporting on these. Note this is only intended to work 1398 acpi_video_switch_brightness(struct work_struct *work) argument 1400 struct acpi_video_device *device = container_of(to_delayed_work(work), 1501 * There are systems where video module known to work fine regardless 1990 "if the current driver doesn't work.\n");
|
H A D | ec.c | 166 struct work_struct work; member in struct:acpi_ec_query 173 static void acpi_ec_event_handler(struct work_struct *work); 465 * more, so it is not necessary to queue up the event work to start the 472 queue_work(ec_wq, &ec->work); 543 flush_workqueue(ec_wq); /* flush ec->work */ 1125 static void acpi_ec_event_processor(struct work_struct *work) argument 1127 struct acpi_ec_query *q = container_of(work, struct acpi_ec_query, work); 1157 INIT_WORK(&q->work, acpi_ec_event_processor); 1208 queue_work(ec_query_wq, &q->work); 1220 acpi_ec_event_handler(struct work_struct *work) argument [all...] |
H A D | internal.h | 82 bool acpi_queue_hotplug_work(struct work_struct *work); 182 EC_EVENT_READY = 0, /* Event work can be submitted */ 183 EC_EVENT_IN_PROGRESS, /* Event work is pending or being processed */ 184 EC_EVENT_COMPLETE, /* Event work processing has completed */ 202 struct work_struct work; member in struct:acpi_ec
|
H A D | osl.c | 48 struct work_struct work; member in struct:acpi_os_dpc 382 static void acpi_os_map_remove(struct work_struct *work) argument 384 struct acpi_ioremap *map = container_of(to_rcu_work(work), 843 static void acpi_os_execute_deferred(struct work_struct *work) argument 845 struct acpi_os_dpc *dpc = container_of(work, struct acpi_os_dpc, work); 1094 INIT_WORK(&dpc->work, acpi_os_execute_deferred); 1103 ret = queue_work(kacpi_notify_wq, &dpc->work); 1113 ret = queue_work_on(0, kacpid_wq, &dpc->work); 1120 pr_err("Unable to queue work\ 1146 struct work_struct work; member in struct:acpi_hp_work 1151 acpi_hotplug_work_fn(struct work_struct *work) argument 1188 acpi_queue_hotplug_work(struct work_struct *work) argument [all...] |
H A D | scan.c | 575 static DECLARE_WORK(work, acpi_device_del_work_fn); 581 * Use the ACPI hotplug workqueue which is ordered, so this work item 582 * won't run after any hotplug work items submitted subsequently. That 586 * run after all of the work items submitted previously, which helps 587 * those work items to ensure that they are not accessing stale device 591 acpi_queue_hotplug_work(&work); 2333 struct work_struct work; member in struct:acpi_scan_clear_dep_work 2337 static void acpi_scan_clear_dep_fn(struct work_struct *work) argument 2341 cdw = container_of(work, struct acpi_scan_clear_dep_work, work); 2772 acpi_table_events_fn(struct work_struct *work) argument 2783 struct work_struct *work; local [all...] |
/linux-master/drivers/acpi/apei/ |
H A D | ghes.c | 166 struct work_struct work; member in struct:ghes_vendor_record_entry 471 * Ensure any queued work has been done before we return to the context that 643 static void ghes_vendor_record_work_func(struct work_struct *work) argument 649 entry = container_of(work, struct ghes_vendor_record_entry, work); 675 INIT_WORK(&entry->work, ghes_vendor_record_work_func); 676 schedule_work(&entry->work); 720 int cxl_cper_register_work(struct work_struct *work) argument 726 cxl_cper_work = work; 731 int cxl_cper_unregister_work(struct work_struct *work) argument [all...] |
/linux-master/drivers/acpi/nfit/ |
H A D | core.c | 2912 static void acpi_nfit_scrub(struct work_struct *work) argument 2918 acpi_desc = container_of(work, typeof(*acpi_desc), dwork.work); 3175 if (work_busy(&acpi_desc->dwork.work))
|