/linux-master/drivers/gpu/drm/ |
H A D | drm_vblank_work.c | 38 * generic delayed work implementation which delays work execution until a 39 * particular vblank has passed, and then executes the work at realtime 43 * re-arming work items can be easily implemented. 48 struct drm_vblank_work *work, *next; local 54 list_for_each_entry_safe(work, next, &vblank->pending_work, node) { 55 if (!drm_vblank_passed(count, work->count)) 58 list_del_init(&work->node); 60 kthread_queue_work(vblank->worker, &work->base); 67 /* Handle cancelling any pending vblank work item 72 struct drm_vblank_work *work, *next; local 109 drm_vblank_work_schedule(struct drm_vblank_work *work, u64 count, bool nextonmiss) argument 185 drm_vblank_work_cancel_sync(struct drm_vblank_work *work) argument 221 drm_vblank_work_flush(struct drm_vblank_work *work) argument 243 drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc, void (*func)(struct kthread_work *work)) argument [all...] |
/linux-master/drivers/media/pci/cobalt/ |
H A D | cobalt-irq.h | 12 void cobalt_irq_work_handler(struct work_struct *work);
|
/linux-master/drivers/net/wireless/st/cw1200/ |
H A D | scan.h | 23 struct work_struct work; member in struct:cw1200_scan 42 void cw1200_scan_work(struct work_struct *work); 43 void cw1200_scan_timeout(struct work_struct *work); 44 void cw1200_clear_recent_scan_work(struct work_struct *work); 51 void cw1200_probe_work(struct work_struct *work);
|
H A D | sta.h | 59 void cw1200_event_handler(struct work_struct *work); 60 void cw1200_bss_loss_work(struct work_struct *work); 61 void cw1200_bss_params_work(struct work_struct *work); 62 void cw1200_keep_alive_work(struct work_struct *work); 63 void cw1200_tx_failure_work(struct work_struct *work); 79 void cw1200_join_timeout(struct work_struct *work); 80 void cw1200_unjoin_work(struct work_struct *work); 81 void cw1200_join_complete_work(struct work_struct *work); 82 void cw1200_wep_key_work(struct work_struct *work); 85 void cw1200_update_filtering_work(struct work_struct *work); [all...] |
/linux-master/drivers/s390/char/ |
H A D | ctrlchar.c | 21 ctrlchar_handle_sysrq(struct work_struct *work) argument 23 struct sysrq_work *sysrq = container_of(work, struct sysrq_work, work); 30 INIT_WORK(&sw->work, ctrlchar_handle_sysrq); 31 schedule_work(&sw->work);
|
/linux-master/net/wireless/ |
H A D | debugfs.c | 115 struct wiphy_work work; member in struct:debugfs_read_work 131 struct wiphy_work *work) 133 struct debugfs_read_work *w = container_of(work, typeof(*w), work); 144 wiphy_work_cancel(w->wiphy, &w->work); 159 struct debugfs_read_work work = { local 167 .completion = COMPLETION_INITIALIZER_ONSTACK(work.completion), 171 .cancel_data = &work, 177 wiphy_work_init(&work.work, wiphy_locked_debugfs_read_wor 130 wiphy_locked_debugfs_read_work(struct wiphy *wiphy, struct wiphy_work *work) argument 195 struct wiphy_work work; member in struct:debugfs_write_work 210 wiphy_locked_debugfs_write_work(struct wiphy *wiphy, struct wiphy_work *work) argument 238 struct debugfs_write_work work = { local [all...] |
/linux-master/lib/ |
H A D | once.c | 9 struct work_struct work; member in struct:once_work 16 struct once_work *work; local 18 work = container_of(w, struct once_work, work); 19 BUG_ON(!static_key_enabled(work->key)); 20 static_branch_disable(work->key); 21 module_put(work->module); 22 kfree(work); 33 INIT_WORK(&w->work, once_deferred); 37 schedule_work(&w->work); [all...] |
/linux-master/include/linux/ |
H A D | kthread.h | 108 * Simple work processor based on kthread. 116 typedef void (*kthread_work_func_t)(struct kthread_work *work); 141 struct kthread_work work; member in struct:kthread_delayed_work 145 #define KTHREAD_WORK_INIT(work, fn) { \ 146 .node = LIST_HEAD_INIT((work).node), \ 151 .work = KTHREAD_WORK_INIT((dwork).work, (fn)), \ 156 #define DEFINE_KTHREAD_WORK(work, fn) \ 157 struct kthread_work work = KTHREAD_WORK_INIT(work, f [all...] |
H A D | workqueue.h | 3 * workqueue.h --- work queue handling for Linux. 20 * The first word is the work queue pointer and the flags rolled into 23 #define work_data_bits(work) ((unsigned long *)(&(work)->data)) 26 WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */ 27 WORK_STRUCT_INACTIVE_BIT, /* work item is inactive */ 29 WORK_STRUCT_LINKED_BIT, /* next work is linked to this one */ 63 * When a work item is off queue, the high bits encode off-queue flags 109 struct work_struct work; member in struct:delayed_work 112 /* target workqueue and CPU ->timer uses to queue ->work */ 118 struct work_struct work; member in struct:rcu_work 203 to_delayed_work(struct work_struct *work) argument 208 to_rcu_work(struct work_struct *work) argument 214 struct work_struct work; member in struct:execute_work 255 work_static(struct work_struct *work) argument 260 __init_work(struct work_struct *work, int onstack) argument 261 destroy_work_on_stack(struct work_struct *work) argument 262 destroy_delayed_work_on_stack(struct delayed_work *work) argument 263 work_static(struct work_struct *work) argument 602 queue_work(struct workqueue_struct *wq, struct work_struct *work) argument 645 schedule_work_on(int cpu, struct work_struct *work) argument 664 schedule_work(struct work_struct *work) argument [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_dim.c | 14 void bnxt_dim_work(struct work_struct *work) argument 16 struct dim *dim = container_of(work, struct dim, work);
|
/linux-master/drivers/staging/octeon/ |
H A D | ethernet-rx.c | 59 * @work: Work queue entry pointing to the packet. 63 static inline int cvm_oct_check_rcv_error(struct cvmx_wqe *work) argument 68 port = work->word0.pip.cn68xx.pknd; 70 port = work->word1.cn38xx.ipprt; 72 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64)) 81 if (work->word2.snoip.err_code == 5 || 82 work->word2.snoip.err_code == 7) { 99 cvmx_phys_to_ptr(work->packet_ptr.s.addr); 102 while (i < work 142 copy_segments_to_skb(struct cvmx_wqe *work, struct sk_buff *skb) argument 222 struct cvmx_wqe *work; local [all...] |
H A D | ethernet-tx.c | 513 /* Get a work queue entry */ 514 struct cvmx_wqe *work = cvmx_fpa_alloc(CVMX_FPA_WQE_POOL); local 516 if (unlikely(!work)) { 517 printk_ratelimited("%s: Failed to allocate a work queue entry\n", 529 cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 1); 555 * Fill in some of the work queue fields. We may need to add 559 work->word0.pip.cn38xx.hw_chksum = skb->csum; 560 work->word1.len = skb->len; 561 cvmx_wqe_set_port(work, priv->port); 562 cvmx_wqe_set_qos(work, pri [all...] |
/linux-master/drivers/net/wireless/ti/wl1251/ |
H A D | ps.h | 18 void wl1251_elp_work(struct work_struct *work);
|
/linux-master/drivers/net/wireless/realtek/rtlwifi/ |
H A D | ps.h | 13 void rtl_ips_nic_off_wq_callback(struct work_struct *work); 20 void rtl_swlps_wq_callback(struct work_struct *work); 21 void rtl_swlps_rfon_wq_callback(struct work_struct *work); 26 void rtl_lps_change_work_callback(struct work_struct *work);
|
/linux-master/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_main.h | 17 void aq_ndev_schedule_work(struct work_struct *work);
|
/linux-master/drivers/media/pci/cx23885/ |
H A D | cx23885-av.c | 14 void cx23885_av_work_handler(struct work_struct *work) argument 17 container_of(work, struct cx23885_dev, cx25840_work);
|
/linux-master/drivers/md/ |
H A D | dm-cache-background-tracker.h | 17 * The cache policy decides what background work should be performed, 19 * is in charge of performing the work, and does so when it sees fit. 22 * work that the policy has decided upon, and handing (issuing) it to 39 * Destroy the tracker. No issued, but not complete, work should 41 * work. 49 * Queue some work within the tracker. 'work' should point to the work 52 * copy of the work. 54 * returns -EINVAL iff the work i [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_dim.c | 44 void mlx5e_rx_dim_work(struct work_struct *work) argument 46 struct dim *dim = container_of(work, struct dim, work); 54 void mlx5e_tx_dim_work(struct work_struct *work) argument 56 struct dim *dim = container_of(work, struct dim, work);
|
/linux-master/kernel/ |
H A D | kthread.c | 307 * functions which do some additional work in non-modular code such as 795 * when they finish. There is defined a safe point for freezing when one work 804 struct kthread_work *work; local 827 work = NULL; 830 work = list_first_entry(&worker->work_list, 832 list_del_init(&work->node); 834 worker->current_work = work; 837 if (work) { 838 kthread_work_func_t func = work->func; 840 trace_sched_kthread_work_execute_start(work); 969 queuing_blocked(struct kthread_worker *worker, struct kthread_work *work) argument 977 kthread_insert_work_sanity_check(struct kthread_worker *worker, struct kthread_work *work) argument 987 kthread_insert_work(struct kthread_worker *worker, struct kthread_work *work, struct list_head *pos) argument 1013 kthread_queue_work(struct kthread_worker *worker, struct kthread_work *work) argument 1040 struct kthread_work *work = &dwork->work; local 1070 struct kthread_work *work = &dwork->work; local 1113 struct kthread_work *work = &dwork->work; local 1130 struct kthread_work work; member in struct:kthread_flush_work 1134 kthread_flush_work_fn(struct kthread_work *work) argument 1147 kthread_flush_work(struct kthread_work *work) argument 1186 kthread_cancel_delayed_work_timer(struct kthread_work *work, unsigned long *flags) argument 1219 __kthread_cancel_work(struct kthread_work *work) argument 1260 struct kthread_work *work = &dwork->work; local 1303 __kthread_cancel_work_sync(struct kthread_work *work, bool is_dwork) argument 1356 kthread_cancel_work_sync(struct kthread_work *work) argument [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_sw_fence_work.c | 16 static void fence_work(struct work_struct *work) argument 18 struct dma_fence_work *f = container_of(work, typeof(*f), work); 20 f->ops->work(f); 39 fence_work(&f->work); 41 queue_work(system_unbound_wq, &f->work); 64 return f->ops->name ?: "work"; 90 INIT_WORK(&f->work, fence_work);
|
/linux-master/tools/perf/ |
H A D | builtin-kwork.c | 314 struct kwork_work *work; local 318 work = container_of(node, struct kwork_work, node); 319 cmp = work_cmp(sort_list, key, work); 325 if (work->name == NULL) 326 work->name = key->name; 327 return work; 361 struct kwork_work *work = zalloc(sizeof(*work)); local 363 if (work == NULL) { 364 pr_err("Failed to zalloc kwork work\ 382 struct kwork_work *work = work_search(root, key, sort_list); local 407 profile_name_match(struct perf_kwork *kwork, struct kwork_work *work) argument 418 profile_event_match(struct perf_kwork *kwork, struct kwork_work *work, struct perf_sample *sample) argument 457 struct kwork_work *work, key; local 515 struct kwork_work *work, key; local 550 struct kwork_work *work; local 578 report_update_exit_event(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument 617 struct kwork_work *work = NULL; local 633 latency_update_entry_event(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument 672 struct kwork_work *work = NULL; local 737 timehist_print_event(struct perf_kwork *kwork, struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample, struct addr_location *al) argument 831 struct kwork_work *work = NULL; local 852 struct kwork_work *work = NULL; local 882 top_update_runtime(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument 913 struct kwork_work *work, *sched_work; local 944 struct kwork_work *work; local 1004 irq_work_init(struct perf_kwork *kwork, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument 1024 irq_work_name(struct kwork_work *work, char *buf, int len) argument 1131 softirq_work_init(struct perf_kwork *kwork, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument 1154 softirq_work_name(struct kwork_work *work, char *buf, int len) argument 1231 workqueue_work_init(struct perf_kwork *kwork __maybe_unused, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine) argument 1250 workqueue_work_name(struct kwork_work *work, char *buf, int len) argument 1299 sched_work_init(struct perf_kwork *kwork __maybe_unused, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument 1319 sched_work_name(struct kwork_work *work, char *buf, int len) argument 1346 report_print_work(struct perf_kwork *kwork, struct kwork_work *work) argument 1637 top_print_work(struct perf_kwork *kwork __maybe_unused, struct kwork_work *work) argument 1836 process_skipped_events(struct perf_kwork *kwork, struct kwork_work *work) argument 1853 struct kwork_work *work = NULL; local 1905 struct kwork_work *work; local 1998 struct kwork_work *work; local 2016 top_calc_idle_time(struct perf_kwork *kwork, struct kwork_work *work) argument 2027 top_calc_irq_runtime(struct perf_kwork *kwork, enum kwork_class_type type, struct kwork_work *work) argument 2042 top_subtract_irq_runtime(struct perf_kwork *kwork, struct kwork_work *work) argument 2070 struct kwork_work *work; local 2098 top_calc_load_runtime(struct perf_kwork *kwork, struct kwork_work *work) argument 2150 struct kwork_work *work; local [all...] |
/linux-master/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_vlan.c | 211 struct work_struct work; member in struct:ipoib_vlan_delete_work 223 * from the global work queue where we can obtain the rtnl_lock safely. 225 static void ipoib_vlan_delete_task(struct work_struct *work) argument 228 container_of(work, struct ipoib_vlan_delete_work, work); 269 struct ipoib_vlan_delete_work *work; local 271 work = kmalloc(sizeof(*work), GFP_KERNEL); 272 if (!work) { 280 work [all...] |
/linux-master/arch/powerpc/platforms/cell/ |
H A D | cpufreq_spudemand.c | 26 struct delayed_work work; member in struct:spu_gov_info_struct 46 static void spu_gov_work(struct work_struct *work) argument 52 info = container_of(work, struct spu_gov_info_struct, work.work); 61 schedule_delayed_work_on(info->policy->cpu, &info->work, delay); 67 INIT_DEFERRABLE_WORK(&info->work, spu_gov_work); 68 schedule_delayed_work_on(info->policy->cpu, &info->work, delay); 73 cancel_delayed_work_sync(&info->work);
|
/linux-master/drivers/extcon/ |
H A D | extcon-gpio.c | 26 * @work: Work fired by the interrupt. 37 struct delayed_work work; member in struct:gpio_extcon_data 45 static void gpio_extcon_work(struct work_struct *work) argument 49 container_of(to_delayed_work(work), struct gpio_extcon_data, 50 work); 60 queue_delayed_work(system_power_efficient_wq, &data->work, 116 ret = devm_delayed_work_autocancel(dev, &data->work, gpio_extcon_work); 132 gpio_extcon_work(&data->work.work); 145 &data->work, dat [all...] |
/linux-master/drivers/input/misc/ |
H A D | da9055_onkey.c | 20 struct delayed_work work; member in struct:da9055_onkey 44 * Hence the deassertion of the pin is simulated through work queue. 47 schedule_delayed_work(&onkey->work, msecs_to_jiffies(10)); 51 static void da9055_onkey_work(struct work_struct *work) argument 53 struct da9055_onkey *onkey = container_of(work, struct da9055_onkey, 54 work.work); 103 INIT_DELAYED_WORK(&onkey->work, da9055_onkey_work); 128 cancel_delayed_work_sync(&onkey->work); 142 cancel_delayed_work_sync(&onkey->work); [all...] |