Searched refs:work (Results 1 - 25 of 3011) sorted by relevance

1234567891011>>

/linux-master/fs/smb/server/
H A Dksmbd_work.c21 struct ksmbd_work *work = kmem_cache_zalloc(work_cache, GFP_KERNEL); local
23 if (work) {
24 work->compound_fid = KSMBD_NO_FID;
25 work->compound_pfid = KSMBD_NO_FID;
26 INIT_LIST_HEAD(&work->request_entry);
27 INIT_LIST_HEAD(&work->async_request_entry);
28 INIT_LIST_HEAD(&work->fp_entry);
29 INIT_LIST_HEAD(&work->interim_entry);
30 INIT_LIST_HEAD(&work->aux_read_list);
31 work
42 ksmbd_free_work_struct(struct ksmbd_work *work) argument
96 ksmbd_queue_work(struct ksmbd_work *work) argument
101 __ksmbd_iov_pin(struct ksmbd_work *work, void *ib, unsigned int ib_len) argument
109 __ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len, void *aux_buf, unsigned int aux_size) argument
159 ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len) argument
164 ksmbd_iov_pin_rsp_read(struct ksmbd_work *work, void *ib, int len, void *aux_buf, unsigned int aux_size) argument
170 allocate_interim_rsp_buf(struct ksmbd_work *work) argument
[all...]
H A Dksmbd_work.h86 struct work_struct work; member in struct:ksmbd_work
97 * @work: smb work containing response buffer
99 static inline void *ksmbd_resp_buf_next(struct ksmbd_work *work) argument
101 return work->response_buf + work->next_smb2_rsp_hdr_off + 4;
106 * @work: smb work containing response buffer
108 static inline void *ksmbd_resp_buf_curr(struct ksmbd_work *work) argument
110 return work
117 ksmbd_req_buf_next(struct ksmbd_work *work) argument
[all...]
H A Dserver.c88 * @work: smb work containing server thread information
92 static inline int check_conn_state(struct ksmbd_work *work) argument
96 if (ksmbd_conn_exiting(work->conn) ||
97 ksmbd_conn_need_reconnect(work->conn)) {
98 rsp_hdr = work->response_buf;
108 static int __process_request(struct ksmbd_work *work, struct ksmbd_conn *conn, argument
115 if (check_conn_state(work))
118 if (ksmbd_verify_smb_message(work)) {
119 conn->ops->set_rsp_status(work, STATUS_INVALID_PARAMETE
163 __handle_ksmbd_work(struct ksmbd_work *work, struct ksmbd_conn *conn) argument
265 struct ksmbd_work *work = container_of(wk, struct ksmbd_work, work); local
292 struct ksmbd_work *work; local
388 server_ctrl_handle_work(struct work_struct *work) argument
[all...]
H A Dsmb2pdu.h435 bool is_smb2_neg_cmd(struct ksmbd_work *work);
436 bool is_smb2_rsp(struct ksmbd_work *work);
438 u16 get_smb2_cmd_val(struct ksmbd_work *work);
439 void set_smb2_rsp_status(struct ksmbd_work *work, __le32 err);
440 int init_smb2_rsp_hdr(struct ksmbd_work *work);
441 int smb2_allocate_rsp_buf(struct ksmbd_work *work);
442 bool is_chained_smb2_message(struct ksmbd_work *work);
443 int init_smb2_neg_rsp(struct ksmbd_work *work);
444 void smb2_set_err_rsp(struct ksmbd_work *work);
445 int smb2_check_user_session(struct ksmbd_work *work);
[all...]
/linux-master/drivers/media/pci/cx23885/
H A Dcx23885-av.h12 void cx23885_av_work_handler(struct work_struct *work);
H A Dcx23885-ir.h15 void cx23885_ir_rx_work_handler(struct work_struct *work);
16 void cx23885_ir_tx_work_handler(struct work_struct *work);
/linux-master/include/linux/
H A Dcompletion.h35 #define COMPLETION_INITIALIZER(work) \
36 { 0, __SWAIT_QUEUE_HEAD_INITIALIZER((work).wait) }
38 #define COMPLETION_INITIALIZER_ONSTACK_MAP(work, map) \
39 (*({ init_completion_map(&(work), &(map)); &(work); }))
41 #define COMPLETION_INITIALIZER_ONSTACK(work) \
42 (*({ init_completion(&work); &work; }))
46 * @work: identifier for the completion structure
52 #define DECLARE_COMPLETION(work) \
[all...]
H A Dirq_work.h37 void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) argument
39 *work = IRQ_WORK_INIT(func);
42 static inline bool irq_work_is_pending(struct irq_work *work) argument
44 return atomic_read(&work->node.a_flags) & IRQ_WORK_PENDING;
47 static inline bool irq_work_is_busy(struct irq_work *work) argument
49 return atomic_read(&work->node.a_flags) & IRQ_WORK_BUSY;
52 static inline bool irq_work_is_hard(struct irq_work *work) argument
54 return atomic_read(&work->node.a_flags) & IRQ_WORK_HARD_IRQ;
57 bool irq_work_queue(struct irq_work *work);
58 bool irq_work_queue_on(struct irq_work *work, in
[all...]
H A Djump_label_ratelimit.h12 struct delayed_work work; member in struct:static_key_deferred
18 struct delayed_work work; member in struct:static_key_true_deferred
24 struct delayed_work work; member in struct:static_key_false_deferred
28 __static_key_slow_dec_deferred(&(x)->key, &(x)->work, (x)->timeout)
30 __static_key_slow_dec_deferred(&(x)->key.key, &(x)->work, (x)->timeout)
33 __static_key_deferred_flush((x), &(x)->work)
37 struct delayed_work *work,
39 extern void __static_key_deferred_flush(void *key, struct delayed_work *work);
43 extern void jump_label_update_timeout(struct work_struct *work);
49 .work
[all...]
/linux-master/include/trace/events/
H A Dworkqueue.h14 * workqueue_queue_work - called when a work gets queued
17 * @work: pointer to struct work_struct
19 * This event occurs when a work is queued immediately or once a
20 * delayed work is actually queued on a workqueue (ie: once the delay
26 struct work_struct *work),
28 TP_ARGS(req_cpu, pwq, work),
31 __field( void *, work )
39 __entry->work = work;
40 __entry->function = work
[all...]
/linux-master/drivers/media/platform/amphion/
H A Dvpu_msgs.h10 void vpu_inst_run_work(struct work_struct *work);
11 void vpu_msg_run_work(struct work_struct *work);
12 void vpu_msg_delayed_work(struct work_struct *work);
/linux-master/drivers/gpu/drm/
H A Ddrm_flip_work.c46 static void drm_flip_work_queue_task(struct drm_flip_work *work, struct drm_flip_task *task) argument
50 spin_lock_irqsave(&work->lock, flags);
51 list_add_tail(&task->node, &work->queued);
52 spin_unlock_irqrestore(&work->lock, flags);
56 * drm_flip_work_queue - queue work
57 * @work: the flip-work
60 * Queues work, that will later be run (passed back to drm_flip_func_t
61 * func) on a work queue after drm_flip_work_commit() is called.
63 void drm_flip_work_queue(struct drm_flip_work *work, voi argument
88 drm_flip_work_commit(struct drm_flip_work *work, struct workqueue_struct *wq) argument
103 struct drm_flip_work *work = container_of(w, struct drm_flip_work, worker); local
134 drm_flip_work_init(struct drm_flip_work *work, const char *name, drm_flip_func_t func) argument
153 drm_flip_work_cleanup(struct drm_flip_work *work) argument
[all...]
/linux-master/kernel/bpf/
H A Dmmap_unlock_work.h26 struct mmap_unlock_irq_work *work = NULL; local
31 work = this_cpu_ptr(&mmap_unlock_work);
32 if (irq_work_is_busy(&work->irq_work)) {
45 *work_ptr = work;
49 static inline void bpf_mmap_unlock_mm(struct mmap_unlock_irq_work *work, struct mm_struct *mm) argument
51 if (!work) {
54 work->mm = mm;
61 irq_work_queue(&work->irq_work);
/linux-master/kernel/
H A Dtask_work.c9 * task_work_add - ask the @task to execute @work->func()
11 * @work: the callback to run
14 * Queue @work for task_work_run() below and notify the @task if @notify
25 * @TWA_RESUME work is run only when the task exits the kernel and returns to
28 * Fails if the @task is exiting/exited and thus it can't process this @work.
29 * Otherwise @work->func() will be called when the @task goes through one of
32 * If the targeted task is exiting, then an error is returned and the work item
42 int task_work_add(struct task_struct *task, struct callback_head *work, argument
47 /* record the work call stack in order to print it in KASAN reports */
48 kasan_record_aux_stack(work);
92 struct callback_head *work; local
150 struct callback_head *work, *head, *next; local
[all...]
H A Dirq_work.c57 static bool irq_work_claim(struct irq_work *work) argument
61 oflags = atomic_fetch_or(IRQ_WORK_CLAIMED | CSD_TYPE_IRQ_WORK, &work->node.a_flags);
63 * If the work is already pending, no need to raise the IPI.
79 static __always_inline void irq_work_raise(struct irq_work *work) argument
82 trace_ipi_send_cpu(smp_processor_id(), _RET_IP_, work->func);
87 /* Enqueue on current CPU, work must already be claimed and preempt disabled */
88 static void __irq_work_queue_local(struct irq_work *work) argument
95 work_flags = atomic_read(&work->node.a_flags);
107 if (!llist_add(&work->node.llist, list))
110 /* If the work i
116 irq_work_queue(struct irq_work *work) argument
137 irq_work_queue_on(struct irq_work *work, int cpu) argument
203 struct irq_work *work = arg; local
237 struct irq_work *work, *tmp; local
286 irq_work_sync(struct irq_work *work) argument
[all...]
/linux-master/tools/testing/selftests/bpf/progs/
H A Dexhandler_kern.c22 struct callback_head *work; local
29 * If task work has been added,
32 * must be specified for the task work.
38 work = task->task_works;
39 func = work->func;
44 barrier_var(work);
45 if (work)
/linux-master/drivers/edac/
H A Dwq.c6 bool edac_queue_work(struct delayed_work *work, unsigned long delay) argument
8 return queue_delayed_work(wq, work, delay);
12 bool edac_mod_work(struct delayed_work *work, unsigned long delay) argument
14 return mod_delayed_work(wq, work, delay);
18 bool edac_stop_work(struct delayed_work *work) argument
22 ret = cancel_delayed_work_sync(work);
/linux-master/virt/kvm/
H A Dasync_pf.c45 static void async_pf_execute(struct work_struct *work) argument
48 container_of(work, struct kvm_async_pf, work);
63 * work item is fully processed.
100 static void kvm_flush_and_free_async_pf_work(struct kvm_async_pf *work) argument
103 * The async #PF is "done", but KVM must wait for the work item itself,
106 * after the last call to module_put(). Note, flushing the work item
112 * need to be flushed (but sanity check that the work wasn't queued).
114 if (work->wakeup_all)
115 WARN_ON_ONCE(work
127 struct kvm_async_pf *work = local
150 struct kvm_async_pf *work = local
166 struct kvm_async_pf *work; local
193 struct kvm_async_pf *work; local
229 struct kvm_async_pf *work; local
[all...]
/linux-master/samples/livepatch/
H A Dlivepatch-callbacks-busymod.c34 static void busymod_work_func(struct work_struct *work);
35 static DECLARE_DELAYED_WORK(work, busymod_work_func);
37 static void busymod_work_func(struct work_struct *work) argument
47 schedule_delayed_work(&work,
54 cancel_delayed_work_sync(&work);
/linux-master/include/drm/
H A Ddrm_vblank_work.h11 * struct drm_vblank_work - A delayed work item which delays until a target
31 * @vblank: A pointer to &drm_vblank_crtc this work item belongs to.
36 * @count: The target vblank this work will execute on. Drivers should
44 * are currently running. A work item cannot be rescheduled until all
50 * @node: The position of this work item in
64 int drm_vblank_work_schedule(struct drm_vblank_work *work,
66 void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
67 void (*func)(struct kthread_work *work));
68 bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work);
69 void drm_vblank_work_flush(struct drm_vblank_work *work);
[all...]
H A Ddrm_flip_work.h34 * Utility to queue up work to run from work-queue context after flip/vblank.
45 * @work: the flip work
48 * Callback function to be called for each of the queue'd work items after
51 typedef void (*drm_flip_func_t)(struct drm_flip_work *work, void *val);
54 * struct drm_flip_work - flip work queue
71 void drm_flip_work_queue(struct drm_flip_work *work, void *val);
72 void drm_flip_work_commit(struct drm_flip_work *work,
74 void drm_flip_work_init(struct drm_flip_work *work,
[all...]
/linux-master/tools/testing/selftests/livepatch/test_modules/
H A Dtest_klp_callbacks_busy.c17 static void busymod_work_func(struct work_struct *work);
18 static DECLARE_WORK(work, busymod_work_func);
21 static void busymod_work_func(struct work_struct *work) argument
40 schedule_work(&work);
44 * exiting until the work function's entry message has printed.
50 * Serialize output: print all messages from the work
53 flush_work(&work);
62 flush_work(&work);
/linux-master/drivers/net/wireless/quantenna/qtnfmac/
H A Devent.h12 void qtnf_event_work_handler(struct work_struct *work);
/linux-master/fs/btrfs/
H A Dasync-thread.c29 /* List head pointing to ordered work list */
55 struct btrfs_fs_info * __pure btrfs_work_owner(const struct btrfs_work *work) argument
57 return work->wq->fs_info;
163 * Hook for threshold which will be called before executing the work,
214 struct btrfs_work *work; local
223 work = list_entry(list->next, struct btrfs_work,
225 if (!test_bit(WORK_DONE_BIT, &work->flags))
231 * updates from ordinary work function.
237 * we leave the work item on the list as a barrier so
238 * that later work item
296 struct btrfs_work *work = container_of(normal_work, struct btrfs_work, local
331 btrfs_init_work(struct btrfs_work *work, btrfs_func_t func, btrfs_ordered_func_t ordered_func) argument
341 btrfs_queue_work(struct btrfs_workqueue *wq, struct btrfs_work *work) argument
[all...]
/linux-master/drivers/accessibility/speakup/
H A Dselection.c20 struct work_struct work; member in struct:speakup_selection_work
25 static void __speakup_set_selection(struct work_struct *work) argument
28 container_of(work, struct speakup_selection_work, work);
58 .work = __WORK_INITIALIZER(speakup_sel_work.work,
65 * cancelling selection work. getting kref first establishes the
87 schedule_work_on(WORK_CPU_UNBOUND, &speakup_sel_work.work);
96 cancel_work_sync(&speakup_sel_work.work);
97 /* setting to null so that if work fail
106 __speakup_paste_selection(struct work_struct *work) argument
[all...]

Completed in 913 milliseconds

1234567891011>>