/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_io_utils.c | 120 kthread_queue_work(work->worker, &work->work); 133 struct kthread_worker *worker, 140 work->worker = worker; 132 msm_hrtimer_work_init(struct msm_hrtimer_work *work, struct kthread_worker *worker, kthread_work_func_t fn, clockid_t clock_id, enum hrtimer_mode mode) argument
|
H A D | msm_kms.c | 213 /* clean up event worker threads */ 215 if (priv->event_thread[i].worker) 216 kthread_destroy_worker(priv->event_thread[i].worker); 272 ev_thread->worker = kthread_create_worker(0, "crtc_event:%d", crtc->base.id); 273 if (IS_ERR(ev_thread->worker)) { 274 ret = PTR_ERR(ev_thread->worker); 276 ev_thread->worker = NULL; 280 sched_set_fifo(ev_thread->worker->task);
|
H A D | msm_fence.c | 24 kthread_queue_work(fctx2gpu(fctx)->worker, &fctx->deadline_work); 157 kthread_queue_work(fctx2gpu(fctx)->worker,
|
/linux-master/drivers/md/ |
H A D | dm-delay.c | 35 struct task_struct *worker; member in struct:delay_c 73 return !!dc->worker; 166 if (dc->worker) 167 kthread_stop(dc->worker); 270 dc->worker = kthread_create(&flush_worker_fn, dc, 271 "dm-delay-flush-worker"); 272 if (IS_ERR(dc->worker)) { 273 ret = PTR_ERR(dc->worker); 274 dc->worker = NULL; 322 wake_up_process(dc->worker); [all...] |
/linux-master/drivers/crypto/caam/ |
H A D | caamrng.c | 43 struct work_struct worker; member in struct:caam_rng_ctx 143 worker); 161 schedule_work(&ctx->worker); 170 flush_work(&ctx->worker); 239 INIT_WORK(&ctx->worker, caam_rng_worker);
|
/linux-master/drivers/block/ |
H A D | loop.c | 819 struct loop_worker *cur_worker, *worker = NULL; local 834 worker = cur_worker; 842 if (worker) 845 worker = kzalloc(sizeof(struct loop_worker), GFP_NOWAIT | __GFP_NOWARN); 847 * In the event we cannot allocate a worker, just queue on the 848 * rootcg worker and issue the I/O as the rootcg 850 if (!worker) { 858 worker->blkcg_css = cmd->blkcg_css; 859 css_get(worker->blkcg_css); 860 INIT_WORK(&worker 893 struct loop_worker *pos, *worker; local 1928 loop_process_work(struct loop_worker *worker, struct list_head *cmd_list, struct loop_device *lo) argument 1964 struct loop_worker *worker = local [all...] |
/linux-master/drivers/macintosh/ams/ |
H A D | ams-core.c | 65 schedule_work(&ams_info.worker); 185 INIT_WORK(&ams_info.worker, ams_worker); 211 /* Flush interrupt worker 216 flush_work(&ams_info.worker);
|
H A D | ams.h | 36 /* Interrupt worker */ 37 struct work_struct worker; member in struct:ams
|
/linux-master/drivers/net/wireguard/ |
H A D | device.h | 30 struct multicore_worker __percpu *worker; member in struct:crypt_queue
|
/linux-master/arch/x86/kvm/ |
H A D | i8254.h | 47 struct kthread_worker *worker; member in struct:kvm_pit
|
H A D | i8254.c | 215 kthread_queue_work(pit->worker, &pit->expired); 276 kthread_queue_work(pt->worker, &pt->expired); 304 * the pt->worker work iterm and reinject the missed tick. 684 pit->worker = kthread_create_worker(0, "kvm-pit/%d", pid_nr); 685 if (IS_ERR(pit->worker)) 728 kthread_destroy_worker(pit->worker); 747 kthread_destroy_worker(pit->worker);
|
/linux-master/kernel/ |
H A D | async.c | 338 * current_is_async - is %current an async worker task? 340 * Returns %true if %current is an async worker task. 344 struct worker *worker = current_wq_worker(); local 346 return worker && worker->current_func == async_run_entry_fn;
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_slpc.c | 20 struct kthread_worker *worker; member in struct:slpc_thread 492 threads[i].worker = kthread_create_worker(0, "igt/slpc_parallel:%d", gt->info.id); 494 if (IS_ERR(threads[i].worker)) { 495 ret = PTR_ERR(threads[i].worker); 501 kthread_queue_work(threads[i].worker, &threads[i].work); 507 if (IS_ERR_OR_NULL(threads[i].worker)) 516 kthread_destroy_worker(threads[i].worker);
|
/linux-master/fs/erofs/ |
H A D | zdata.c | 343 struct kthread_worker *worker; local 347 worker = rcu_dereference_protected( 350 if (worker) 351 kthread_destroy_worker(worker); 358 struct kthread_worker *worker = local 361 if (IS_ERR(worker)) 362 return worker; 364 sched_set_fifo_low(worker->task); 365 return worker; 370 struct kthread_worker *worker; local 396 struct kthread_worker *worker, *old; local 415 struct kthread_worker *worker; local 1389 struct kthread_worker *worker; local [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_ct.h | 84 struct work_struct worker; /* handler for incoming requests */ member in struct:intel_guc_ct::__anon675
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_sync.c | 26 struct work_struct worker; member in struct:xe_user_fence 73 struct xe_user_fence *ufence = container_of(w, struct xe_user_fence, worker); 90 INIT_WORK(&ufence->worker, user_fence_worker); 91 queue_work(ufence->xe->ordered_wq, &ufence->worker);
|
H A D | xe_gt_pagefault.c | 351 queue_work(gt->usm.pf_wq, &pf_queue->worker); 364 struct pf_queue *pf_queue = container_of(w, struct pf_queue, worker); 416 INIT_WORK(>->usm.pf_queue[i].worker, pf_queue_work_func); 421 INIT_WORK(>->usm.acc_queue[i].worker, acc_queue_work_func); 597 struct acc_queue *acc_queue = container_of(w, struct acc_queue, worker); 653 queue_work(gt->usm.acc_wq, &acc_queue->worker);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_internal.h | 104 kthread_flush_worker(vblank->worker); 109 if (vblank->worker) 110 kthread_destroy_worker(vblank->worker);
|
/linux-master/drivers/platform/olpc/ |
H A D | olpc-ec.c | 36 struct work_struct worker; member in struct:olpc_ec_priv 79 struct olpc_ec_priv *ec = container_of(w, struct olpc_ec_priv, worker); 104 /* Run the worker thread again in case there are more cmds pending */ 105 schedule_work(&ec->worker); 123 schedule_work(&ec->worker); 160 /* The worker thread dequeues the cmd; no need to do anything here */ 419 INIT_WORK(&ec->worker, olpc_ec_worker);
|
/linux-master/drivers/hid/ |
H A D | hid-bigbenff.c | 185 struct work_struct worker; member in struct:bigben_device 194 schedule_work(&bigben->worker); 201 struct bigben_device, worker); 365 cancel_work_sync(&bigben->worker); 414 INIT_WORK(&bigben->worker, bigben_worker);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_request.c | 303 struct kthread_worker *worker; member in struct:smoke_thread 493 struct kthread_worker *worker; local 495 worker = kthread_create_worker(0, "igt/%d", n); 496 if (IS_ERR(worker)) { 497 ret = PTR_ERR(worker); 502 threads[n].worker = worker; 509 kthread_queue_work(worker, &threads[n].work); 523 kthread_destroy_worker(threads[n].worker); 1458 struct kthread_worker *worker; member in struct:parallel_thread 1646 struct kthread_worker *worker; local 1807 struct kthread_worker *worker; local 2958 struct kthread_worker *worker; member in struct:p_thread 3216 struct kthread_worker *worker; local [all...] |
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_tx.c | 177 queue_work(dpmaif_ctrl->txq[txq->index].worker, 533 queue_work(dpmaif_ctrl->txq[i].worker, &dpmaif_ctrl->txq[i].dpmaif_tx_work); 621 txq->worker = alloc_ordered_workqueue("md_dpmaif_tx%d_worker", 624 if (!txq->worker) 635 if (txq->worker) 636 destroy_workqueue(txq->worker);
|
/linux-master/lib/ |
H A D | test_objpool.c | 61 /* per-cpu worker */ 66 void (*worker)(struct ot_item *item, int irq); member in struct:ot_item 165 item->worker(item, 1); 201 void (*worker)(struct ot_item *, int)) 206 item->worker = worker; 232 item->worker(item, 0); 397 /* yeild cpu to worker threads for duration ms */ 586 /* yeild cpu to worker threads for duration ms */ 605 * by rcu callback or by worker thread [all...] |
/linux-master/include/trace/events/ |
H A D | sched.h | 57 * @worker: pointer to the kthread_worker 66 TP_PROTO(struct kthread_worker *worker, 69 TP_ARGS(worker, work), 74 __field( void *, worker) 80 __entry->worker = worker; 83 TP_printk("work struct=%p function=%ps worker=%p", 84 __entry->work, __entry->function, __entry->worker) 115 * @function: pointer to worker function
|
/linux-master/tools/perf/bench/ |
H A D | futex-wake-parallel.c | 41 pthread_t worker; member in struct:thread_data 115 if (pthread_create(&td[i].worker, &thread_attr, 123 if (pthread_join(td[i].worker, NULL))
|