Searched refs:work (Results 26 - 50 of 3012) sorted by relevance

1234567891011>>

/linux-master/drivers/gpu/drm/
H A Ddrm_vblank_work.c38 * generic delayed work implementation which delays work execution until a
39 * particular vblank has passed, and then executes the work at realtime
43 * re-arming work items can be easily implemented.
48 struct drm_vblank_work *work, *next; local
54 list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
55 if (!drm_vblank_passed(count, work->count))
58 list_del_init(&work->node);
60 kthread_queue_work(vblank->worker, &work->base);
67 /* Handle cancelling any pending vblank work item
72 struct drm_vblank_work *work, *next; local
109 drm_vblank_work_schedule(struct drm_vblank_work *work, u64 count, bool nextonmiss) argument
185 drm_vblank_work_cancel_sync(struct drm_vblank_work *work) argument
221 drm_vblank_work_flush(struct drm_vblank_work *work) argument
243 drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc, void (*func)(struct kthread_work *work)) argument
[all...]
/linux-master/drivers/media/pci/cobalt/
H A Dcobalt-irq.h12 void cobalt_irq_work_handler(struct work_struct *work);
/linux-master/drivers/net/wireless/st/cw1200/
H A Dscan.h23 struct work_struct work; member in struct:cw1200_scan
42 void cw1200_scan_work(struct work_struct *work);
43 void cw1200_scan_timeout(struct work_struct *work);
44 void cw1200_clear_recent_scan_work(struct work_struct *work);
51 void cw1200_probe_work(struct work_struct *work);
H A Dsta.h59 void cw1200_event_handler(struct work_struct *work);
60 void cw1200_bss_loss_work(struct work_struct *work);
61 void cw1200_bss_params_work(struct work_struct *work);
62 void cw1200_keep_alive_work(struct work_struct *work);
63 void cw1200_tx_failure_work(struct work_struct *work);
79 void cw1200_join_timeout(struct work_struct *work);
80 void cw1200_unjoin_work(struct work_struct *work);
81 void cw1200_join_complete_work(struct work_struct *work);
82 void cw1200_wep_key_work(struct work_struct *work);
85 void cw1200_update_filtering_work(struct work_struct *work);
[all...]
/linux-master/drivers/s390/char/
H A Dctrlchar.c21 ctrlchar_handle_sysrq(struct work_struct *work) argument
23 struct sysrq_work *sysrq = container_of(work, struct sysrq_work, work);
30 INIT_WORK(&sw->work, ctrlchar_handle_sysrq);
31 schedule_work(&sw->work);
/linux-master/net/wireless/
H A Ddebugfs.c115 struct wiphy_work work; member in struct:debugfs_read_work
131 struct wiphy_work *work)
133 struct debugfs_read_work *w = container_of(work, typeof(*w), work);
144 wiphy_work_cancel(w->wiphy, &w->work);
159 struct debugfs_read_work work = { local
167 .completion = COMPLETION_INITIALIZER_ONSTACK(work.completion),
171 .cancel_data = &work,
177 wiphy_work_init(&work.work, wiphy_locked_debugfs_read_wor
130 wiphy_locked_debugfs_read_work(struct wiphy *wiphy, struct wiphy_work *work) argument
195 struct wiphy_work work; member in struct:debugfs_write_work
210 wiphy_locked_debugfs_write_work(struct wiphy *wiphy, struct wiphy_work *work) argument
238 struct debugfs_write_work work = { local
[all...]
/linux-master/lib/
H A Donce.c9 struct work_struct work; member in struct:once_work
16 struct once_work *work; local
18 work = container_of(w, struct once_work, work);
19 BUG_ON(!static_key_enabled(work->key));
20 static_branch_disable(work->key);
21 module_put(work->module);
22 kfree(work);
33 INIT_WORK(&w->work, once_deferred);
37 schedule_work(&w->work);
[all...]
/linux-master/include/linux/
H A Dkthread.h108 * Simple work processor based on kthread.
116 typedef void (*kthread_work_func_t)(struct kthread_work *work);
141 struct kthread_work work; member in struct:kthread_delayed_work
145 #define KTHREAD_WORK_INIT(work, fn) { \
146 .node = LIST_HEAD_INIT((work).node), \
151 .work = KTHREAD_WORK_INIT((dwork).work, (fn)), \
156 #define DEFINE_KTHREAD_WORK(work, fn) \
157 struct kthread_work work = KTHREAD_WORK_INIT(work, f
[all...]
H A Dworkqueue.h3 * workqueue.h --- work queue handling for Linux.
20 * The first word is the work queue pointer and the flags rolled into
23 #define work_data_bits(work) ((unsigned long *)(&(work)->data))
26 WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */
27 WORK_STRUCT_INACTIVE_BIT, /* work item is inactive */
29 WORK_STRUCT_LINKED_BIT, /* next work is linked to this one */
63 * When a work item is off queue, the high bits encode off-queue flags
109 struct work_struct work; member in struct:delayed_work
112 /* target workqueue and CPU ->timer uses to queue ->work */
118 struct work_struct work; member in struct:rcu_work
203 to_delayed_work(struct work_struct *work) argument
208 to_rcu_work(struct work_struct *work) argument
214 struct work_struct work; member in struct:execute_work
255 work_static(struct work_struct *work) argument
260 __init_work(struct work_struct *work, int onstack) argument
261 destroy_work_on_stack(struct work_struct *work) argument
262 destroy_delayed_work_on_stack(struct delayed_work *work) argument
263 work_static(struct work_struct *work) argument
602 queue_work(struct workqueue_struct *wq, struct work_struct *work) argument
645 schedule_work_on(int cpu, struct work_struct *work) argument
664 schedule_work(struct work_struct *work) argument
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_dim.c14 void bnxt_dim_work(struct work_struct *work) argument
16 struct dim *dim = container_of(work, struct dim, work);
/linux-master/drivers/staging/octeon/
H A Dethernet-rx.c59 * @work: Work queue entry pointing to the packet.
63 static inline int cvm_oct_check_rcv_error(struct cvmx_wqe *work) argument
68 port = work->word0.pip.cn68xx.pknd;
70 port = work->word1.cn38xx.ipprt;
72 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64))
81 if (work->word2.snoip.err_code == 5 ||
82 work->word2.snoip.err_code == 7) {
99 cvmx_phys_to_ptr(work->packet_ptr.s.addr);
102 while (i < work
142 copy_segments_to_skb(struct cvmx_wqe *work, struct sk_buff *skb) argument
222 struct cvmx_wqe *work; local
[all...]
H A Dethernet-tx.c513 /* Get a work queue entry */
514 struct cvmx_wqe *work = cvmx_fpa_alloc(CVMX_FPA_WQE_POOL); local
516 if (unlikely(!work)) {
517 printk_ratelimited("%s: Failed to allocate a work queue entry\n",
529 cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 1);
555 * Fill in some of the work queue fields. We may need to add
559 work->word0.pip.cn38xx.hw_chksum = skb->csum;
560 work->word1.len = skb->len;
561 cvmx_wqe_set_port(work, priv->port);
562 cvmx_wqe_set_qos(work, pri
[all...]
/linux-master/drivers/net/wireless/ti/wl1251/
H A Dps.h18 void wl1251_elp_work(struct work_struct *work);
/linux-master/drivers/net/wireless/realtek/rtlwifi/
H A Dps.h13 void rtl_ips_nic_off_wq_callback(struct work_struct *work);
20 void rtl_swlps_wq_callback(struct work_struct *work);
21 void rtl_swlps_rfon_wq_callback(struct work_struct *work);
26 void rtl_lps_change_work_callback(struct work_struct *work);
/linux-master/drivers/net/ethernet/aquantia/atlantic/
H A Daq_main.h17 void aq_ndev_schedule_work(struct work_struct *work);
/linux-master/drivers/media/pci/cx23885/
H A Dcx23885-av.c14 void cx23885_av_work_handler(struct work_struct *work) argument
17 container_of(work, struct cx23885_dev, cx25840_work);
/linux-master/drivers/md/
H A Ddm-cache-background-tracker.h17 * The cache policy decides what background work should be performed,
19 * is in charge of performing the work, and does so when it sees fit.
22 * work that the policy has decided upon, and handing (issuing) it to
39 * Destroy the tracker. No issued, but not complete, work should
41 * work.
49 * Queue some work within the tracker. 'work' should point to the work
52 * copy of the work.
54 * returns -EINVAL iff the work i
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_dim.c44 void mlx5e_rx_dim_work(struct work_struct *work) argument
46 struct dim *dim = container_of(work, struct dim, work);
54 void mlx5e_tx_dim_work(struct work_struct *work) argument
56 struct dim *dim = container_of(work, struct dim, work);
/linux-master/kernel/
H A Dkthread.c307 * functions which do some additional work in non-modular code such as
795 * when they finish. There is defined a safe point for freezing when one work
804 struct kthread_work *work; local
827 work = NULL;
830 work = list_first_entry(&worker->work_list,
832 list_del_init(&work->node);
834 worker->current_work = work;
837 if (work) {
838 kthread_work_func_t func = work->func;
840 trace_sched_kthread_work_execute_start(work);
969 queuing_blocked(struct kthread_worker *worker, struct kthread_work *work) argument
977 kthread_insert_work_sanity_check(struct kthread_worker *worker, struct kthread_work *work) argument
987 kthread_insert_work(struct kthread_worker *worker, struct kthread_work *work, struct list_head *pos) argument
1013 kthread_queue_work(struct kthread_worker *worker, struct kthread_work *work) argument
1040 struct kthread_work *work = &dwork->work; local
1070 struct kthread_work *work = &dwork->work; local
1113 struct kthread_work *work = &dwork->work; local
1130 struct kthread_work work; member in struct:kthread_flush_work
1134 kthread_flush_work_fn(struct kthread_work *work) argument
1147 kthread_flush_work(struct kthread_work *work) argument
1186 kthread_cancel_delayed_work_timer(struct kthread_work *work, unsigned long *flags) argument
1219 __kthread_cancel_work(struct kthread_work *work) argument
1260 struct kthread_work *work = &dwork->work; local
1303 __kthread_cancel_work_sync(struct kthread_work *work, bool is_dwork) argument
1356 kthread_cancel_work_sync(struct kthread_work *work) argument
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_sw_fence_work.c16 static void fence_work(struct work_struct *work) argument
18 struct dma_fence_work *f = container_of(work, typeof(*f), work);
20 f->ops->work(f);
39 fence_work(&f->work);
41 queue_work(system_unbound_wq, &f->work);
64 return f->ops->name ?: "work";
90 INIT_WORK(&f->work, fence_work);
/linux-master/tools/perf/
H A Dbuiltin-kwork.c314 struct kwork_work *work; local
318 work = container_of(node, struct kwork_work, node);
319 cmp = work_cmp(sort_list, key, work);
325 if (work->name == NULL)
326 work->name = key->name;
327 return work;
361 struct kwork_work *work = zalloc(sizeof(*work)); local
363 if (work == NULL) {
364 pr_err("Failed to zalloc kwork work\
382 struct kwork_work *work = work_search(root, key, sort_list); local
407 profile_name_match(struct perf_kwork *kwork, struct kwork_work *work) argument
418 profile_event_match(struct perf_kwork *kwork, struct kwork_work *work, struct perf_sample *sample) argument
457 struct kwork_work *work, key; local
515 struct kwork_work *work, key; local
550 struct kwork_work *work; local
578 report_update_exit_event(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument
617 struct kwork_work *work = NULL; local
633 latency_update_entry_event(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument
672 struct kwork_work *work = NULL; local
737 timehist_print_event(struct perf_kwork *kwork, struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample, struct addr_location *al) argument
831 struct kwork_work *work = NULL; local
852 struct kwork_work *work = NULL; local
882 top_update_runtime(struct kwork_work *work, struct kwork_atom *atom, struct perf_sample *sample) argument
913 struct kwork_work *work, *sched_work; local
944 struct kwork_work *work; local
1004 irq_work_init(struct perf_kwork *kwork, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument
1024 irq_work_name(struct kwork_work *work, char *buf, int len) argument
1131 softirq_work_init(struct perf_kwork *kwork, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument
1154 softirq_work_name(struct kwork_work *work, char *buf, int len) argument
1231 workqueue_work_init(struct perf_kwork *kwork __maybe_unused, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type __maybe_unused, struct evsel *evsel, struct perf_sample *sample, struct machine *machine) argument
1250 workqueue_work_name(struct kwork_work *work, char *buf, int len) argument
1299 sched_work_init(struct perf_kwork *kwork __maybe_unused, struct kwork_class *class, struct kwork_work *work, enum kwork_trace_type src_type, struct evsel *evsel, struct perf_sample *sample, struct machine *machine __maybe_unused) argument
1319 sched_work_name(struct kwork_work *work, char *buf, int len) argument
1346 report_print_work(struct perf_kwork *kwork, struct kwork_work *work) argument
1637 top_print_work(struct perf_kwork *kwork __maybe_unused, struct kwork_work *work) argument
1836 process_skipped_events(struct perf_kwork *kwork, struct kwork_work *work) argument
1853 struct kwork_work *work = NULL; local
1905 struct kwork_work *work; local
1998 struct kwork_work *work; local
2016 top_calc_idle_time(struct perf_kwork *kwork, struct kwork_work *work) argument
2027 top_calc_irq_runtime(struct perf_kwork *kwork, enum kwork_class_type type, struct kwork_work *work) argument
2042 top_subtract_irq_runtime(struct perf_kwork *kwork, struct kwork_work *work) argument
2070 struct kwork_work *work; local
2098 top_calc_load_runtime(struct perf_kwork *kwork, struct kwork_work *work) argument
2150 struct kwork_work *work; local
[all...]
/linux-master/drivers/infiniband/ulp/ipoib/
H A Dipoib_vlan.c211 struct work_struct work; member in struct:ipoib_vlan_delete_work
223 * from the global work queue where we can obtain the rtnl_lock safely.
225 static void ipoib_vlan_delete_task(struct work_struct *work) argument
228 container_of(work, struct ipoib_vlan_delete_work, work);
269 struct ipoib_vlan_delete_work *work; local
271 work = kmalloc(sizeof(*work), GFP_KERNEL);
272 if (!work) {
280 work
[all...]
/linux-master/arch/powerpc/platforms/cell/
H A Dcpufreq_spudemand.c26 struct delayed_work work; member in struct:spu_gov_info_struct
46 static void spu_gov_work(struct work_struct *work) argument
52 info = container_of(work, struct spu_gov_info_struct, work.work);
61 schedule_delayed_work_on(info->policy->cpu, &info->work, delay);
67 INIT_DEFERRABLE_WORK(&info->work, spu_gov_work);
68 schedule_delayed_work_on(info->policy->cpu, &info->work, delay);
73 cancel_delayed_work_sync(&info->work);
/linux-master/drivers/extcon/
H A Dextcon-gpio.c26 * @work: Work fired by the interrupt.
37 struct delayed_work work; member in struct:gpio_extcon_data
45 static void gpio_extcon_work(struct work_struct *work) argument
49 container_of(to_delayed_work(work), struct gpio_extcon_data,
50 work);
60 queue_delayed_work(system_power_efficient_wq, &data->work,
116 ret = devm_delayed_work_autocancel(dev, &data->work, gpio_extcon_work);
132 gpio_extcon_work(&data->work.work);
145 &data->work, dat
[all...]
/linux-master/drivers/input/misc/
H A Dda9055_onkey.c20 struct delayed_work work; member in struct:da9055_onkey
44 * Hence the deassertion of the pin is simulated through work queue.
47 schedule_delayed_work(&onkey->work, msecs_to_jiffies(10));
51 static void da9055_onkey_work(struct work_struct *work) argument
53 struct da9055_onkey *onkey = container_of(work, struct da9055_onkey,
54 work.work);
103 INIT_DELAYED_WORK(&onkey->work, da9055_onkey_work);
128 cancel_delayed_work_sync(&onkey->work);
142 cancel_delayed_work_sync(&onkey->work);
[all...]

Completed in 247 milliseconds

1234567891011>>