/linux-master/fs/smb/server/ |
H A D | ksmbd_work.c | 21 struct ksmbd_work *work = kmem_cache_zalloc(work_cache, GFP_KERNEL); local 23 if (work) { 24 work->compound_fid = KSMBD_NO_FID; 25 work->compound_pfid = KSMBD_NO_FID; 26 INIT_LIST_HEAD(&work->request_entry); 27 INIT_LIST_HEAD(&work->async_request_entry); 28 INIT_LIST_HEAD(&work->fp_entry); 29 INIT_LIST_HEAD(&work->interim_entry); 30 INIT_LIST_HEAD(&work->aux_read_list); 31 work 42 ksmbd_free_work_struct(struct ksmbd_work *work) argument 96 ksmbd_queue_work(struct ksmbd_work *work) argument 101 __ksmbd_iov_pin(struct ksmbd_work *work, void *ib, unsigned int ib_len) argument 109 __ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len, void *aux_buf, unsigned int aux_size) argument 159 ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len) argument 164 ksmbd_iov_pin_rsp_read(struct ksmbd_work *work, void *ib, int len, void *aux_buf, unsigned int aux_size) argument 170 allocate_interim_rsp_buf(struct ksmbd_work *work) argument [all...] |
H A D | ksmbd_work.h | 86 struct work_struct work; member in struct:ksmbd_work 97 * @work: smb work containing response buffer 99 static inline void *ksmbd_resp_buf_next(struct ksmbd_work *work) argument 101 return work->response_buf + work->next_smb2_rsp_hdr_off + 4; 106 * @work: smb work containing response buffer 108 static inline void *ksmbd_resp_buf_curr(struct ksmbd_work *work) argument 110 return work 117 ksmbd_req_buf_next(struct ksmbd_work *work) argument [all...] |
H A D | server.c | 88 * @work: smb work containing server thread information 92 static inline int check_conn_state(struct ksmbd_work *work) argument 96 if (ksmbd_conn_exiting(work->conn) || 97 ksmbd_conn_need_reconnect(work->conn)) { 98 rsp_hdr = work->response_buf; 108 static int __process_request(struct ksmbd_work *work, struct ksmbd_conn *conn, argument 115 if (check_conn_state(work)) 118 if (ksmbd_verify_smb_message(work)) { 119 conn->ops->set_rsp_status(work, STATUS_INVALID_PARAMETE 163 __handle_ksmbd_work(struct ksmbd_work *work, struct ksmbd_conn *conn) argument 265 struct ksmbd_work *work = container_of(wk, struct ksmbd_work, work); local 292 struct ksmbd_work *work; local 388 server_ctrl_handle_work(struct work_struct *work) argument [all...] |
H A D | smb2pdu.h | 435 bool is_smb2_neg_cmd(struct ksmbd_work *work); 436 bool is_smb2_rsp(struct ksmbd_work *work); 438 u16 get_smb2_cmd_val(struct ksmbd_work *work); 439 void set_smb2_rsp_status(struct ksmbd_work *work, __le32 err); 440 int init_smb2_rsp_hdr(struct ksmbd_work *work); 441 int smb2_allocate_rsp_buf(struct ksmbd_work *work); 442 bool is_chained_smb2_message(struct ksmbd_work *work); 443 int init_smb2_neg_rsp(struct ksmbd_work *work); 444 void smb2_set_err_rsp(struct ksmbd_work *work); 445 int smb2_check_user_session(struct ksmbd_work *work); [all...] |
/linux-master/drivers/media/pci/cx23885/ |
H A D | cx23885-av.h | 12 void cx23885_av_work_handler(struct work_struct *work);
|
H A D | cx23885-ir.h | 15 void cx23885_ir_rx_work_handler(struct work_struct *work); 16 void cx23885_ir_tx_work_handler(struct work_struct *work);
|
/linux-master/include/linux/ |
H A D | completion.h | 35 #define COMPLETION_INITIALIZER(work) \ 36 { 0, __SWAIT_QUEUE_HEAD_INITIALIZER((work).wait) } 38 #define COMPLETION_INITIALIZER_ONSTACK_MAP(work, map) \ 39 (*({ init_completion_map(&(work), &(map)); &(work); })) 41 #define COMPLETION_INITIALIZER_ONSTACK(work) \ 42 (*({ init_completion(&work); &work; })) 46 * @work: identifier for the completion structure 52 #define DECLARE_COMPLETION(work) \ [all...] |
H A D | irq_work.h | 37 void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) argument 39 *work = IRQ_WORK_INIT(func); 42 static inline bool irq_work_is_pending(struct irq_work *work) argument 44 return atomic_read(&work->node.a_flags) & IRQ_WORK_PENDING; 47 static inline bool irq_work_is_busy(struct irq_work *work) argument 49 return atomic_read(&work->node.a_flags) & IRQ_WORK_BUSY; 52 static inline bool irq_work_is_hard(struct irq_work *work) argument 54 return atomic_read(&work->node.a_flags) & IRQ_WORK_HARD_IRQ; 57 bool irq_work_queue(struct irq_work *work); 58 bool irq_work_queue_on(struct irq_work *work, in [all...] |
H A D | jump_label_ratelimit.h | 12 struct delayed_work work; member in struct:static_key_deferred 18 struct delayed_work work; member in struct:static_key_true_deferred 24 struct delayed_work work; member in struct:static_key_false_deferred 28 __static_key_slow_dec_deferred(&(x)->key, &(x)->work, (x)->timeout) 30 __static_key_slow_dec_deferred(&(x)->key.key, &(x)->work, (x)->timeout) 33 __static_key_deferred_flush((x), &(x)->work) 37 struct delayed_work *work, 39 extern void __static_key_deferred_flush(void *key, struct delayed_work *work); 43 extern void jump_label_update_timeout(struct work_struct *work); 49 .work [all...] |
/linux-master/include/trace/events/ |
H A D | workqueue.h | 14 * workqueue_queue_work - called when a work gets queued 17 * @work: pointer to struct work_struct 19 * This event occurs when a work is queued immediately or once a 20 * delayed work is actually queued on a workqueue (ie: once the delay 26 struct work_struct *work), 28 TP_ARGS(req_cpu, pwq, work), 31 __field( void *, work ) 39 __entry->work = work; 40 __entry->function = work [all...] |
/linux-master/drivers/media/platform/amphion/ |
H A D | vpu_msgs.h | 10 void vpu_inst_run_work(struct work_struct *work); 11 void vpu_msg_run_work(struct work_struct *work); 12 void vpu_msg_delayed_work(struct work_struct *work);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_flip_work.c | 46 static void drm_flip_work_queue_task(struct drm_flip_work *work, struct drm_flip_task *task) argument 50 spin_lock_irqsave(&work->lock, flags); 51 list_add_tail(&task->node, &work->queued); 52 spin_unlock_irqrestore(&work->lock, flags); 56 * drm_flip_work_queue - queue work 57 * @work: the flip-work 60 * Queues work, that will later be run (passed back to drm_flip_func_t 61 * func) on a work queue after drm_flip_work_commit() is called. 63 void drm_flip_work_queue(struct drm_flip_work *work, voi argument 88 drm_flip_work_commit(struct drm_flip_work *work, struct workqueue_struct *wq) argument 103 struct drm_flip_work *work = container_of(w, struct drm_flip_work, worker); local 134 drm_flip_work_init(struct drm_flip_work *work, const char *name, drm_flip_func_t func) argument 153 drm_flip_work_cleanup(struct drm_flip_work *work) argument [all...] |
/linux-master/kernel/bpf/ |
H A D | mmap_unlock_work.h | 26 struct mmap_unlock_irq_work *work = NULL; local 31 work = this_cpu_ptr(&mmap_unlock_work); 32 if (irq_work_is_busy(&work->irq_work)) { 45 *work_ptr = work; 49 static inline void bpf_mmap_unlock_mm(struct mmap_unlock_irq_work *work, struct mm_struct *mm) argument 51 if (!work) { 54 work->mm = mm; 61 irq_work_queue(&work->irq_work);
|
/linux-master/kernel/ |
H A D | task_work.c | 9 * task_work_add - ask the @task to execute @work->func() 11 * @work: the callback to run 14 * Queue @work for task_work_run() below and notify the @task if @notify 25 * @TWA_RESUME work is run only when the task exits the kernel and returns to 28 * Fails if the @task is exiting/exited and thus it can't process this @work. 29 * Otherwise @work->func() will be called when the @task goes through one of 32 * If the targeted task is exiting, then an error is returned and the work item 42 int task_work_add(struct task_struct *task, struct callback_head *work, argument 47 /* record the work call stack in order to print it in KASAN reports */ 48 kasan_record_aux_stack(work); 92 struct callback_head *work; local 150 struct callback_head *work, *head, *next; local [all...] |
H A D | irq_work.c | 57 static bool irq_work_claim(struct irq_work *work) argument 61 oflags = atomic_fetch_or(IRQ_WORK_CLAIMED | CSD_TYPE_IRQ_WORK, &work->node.a_flags); 63 * If the work is already pending, no need to raise the IPI. 79 static __always_inline void irq_work_raise(struct irq_work *work) argument 82 trace_ipi_send_cpu(smp_processor_id(), _RET_IP_, work->func); 87 /* Enqueue on current CPU, work must already be claimed and preempt disabled */ 88 static void __irq_work_queue_local(struct irq_work *work) argument 95 work_flags = atomic_read(&work->node.a_flags); 107 if (!llist_add(&work->node.llist, list)) 110 /* If the work i 116 irq_work_queue(struct irq_work *work) argument 137 irq_work_queue_on(struct irq_work *work, int cpu) argument 203 struct irq_work *work = arg; local 237 struct irq_work *work, *tmp; local 286 irq_work_sync(struct irq_work *work) argument [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | exhandler_kern.c | 22 struct callback_head *work; local 29 * If task work has been added, 32 * must be specified for the task work. 38 work = task->task_works; 39 func = work->func; 44 barrier_var(work); 45 if (work)
|
/linux-master/drivers/edac/ |
H A D | wq.c | 6 bool edac_queue_work(struct delayed_work *work, unsigned long delay) argument 8 return queue_delayed_work(wq, work, delay); 12 bool edac_mod_work(struct delayed_work *work, unsigned long delay) argument 14 return mod_delayed_work(wq, work, delay); 18 bool edac_stop_work(struct delayed_work *work) argument 22 ret = cancel_delayed_work_sync(work);
|
/linux-master/virt/kvm/ |
H A D | async_pf.c | 45 static void async_pf_execute(struct work_struct *work) argument 48 container_of(work, struct kvm_async_pf, work); 63 * work item is fully processed. 100 static void kvm_flush_and_free_async_pf_work(struct kvm_async_pf *work) argument 103 * The async #PF is "done", but KVM must wait for the work item itself, 106 * after the last call to module_put(). Note, flushing the work item 112 * need to be flushed (but sanity check that the work wasn't queued). 114 if (work->wakeup_all) 115 WARN_ON_ONCE(work 127 struct kvm_async_pf *work = local 150 struct kvm_async_pf *work = local 166 struct kvm_async_pf *work; local 193 struct kvm_async_pf *work; local 229 struct kvm_async_pf *work; local [all...] |
/linux-master/samples/livepatch/ |
H A D | livepatch-callbacks-busymod.c | 34 static void busymod_work_func(struct work_struct *work); 35 static DECLARE_DELAYED_WORK(work, busymod_work_func); 37 static void busymod_work_func(struct work_struct *work) argument 47 schedule_delayed_work(&work, 54 cancel_delayed_work_sync(&work);
|
/linux-master/include/drm/ |
H A D | drm_vblank_work.h | 11 * struct drm_vblank_work - A delayed work item which delays until a target 31 * @vblank: A pointer to &drm_vblank_crtc this work item belongs to. 36 * @count: The target vblank this work will execute on. Drivers should 44 * are currently running. A work item cannot be rescheduled until all 50 * @node: The position of this work item in 64 int drm_vblank_work_schedule(struct drm_vblank_work *work, 66 void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc, 67 void (*func)(struct kthread_work *work)); 68 bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work); 69 void drm_vblank_work_flush(struct drm_vblank_work *work); [all...] |
H A D | drm_flip_work.h | 34 * Utility to queue up work to run from work-queue context after flip/vblank. 45 * @work: the flip work 48 * Callback function to be called for each of the queue'd work items after 51 typedef void (*drm_flip_func_t)(struct drm_flip_work *work, void *val); 54 * struct drm_flip_work - flip work queue 71 void drm_flip_work_queue(struct drm_flip_work *work, void *val); 72 void drm_flip_work_commit(struct drm_flip_work *work, 74 void drm_flip_work_init(struct drm_flip_work *work, [all...] |
/linux-master/tools/testing/selftests/livepatch/test_modules/ |
H A D | test_klp_callbacks_busy.c | 17 static void busymod_work_func(struct work_struct *work); 18 static DECLARE_WORK(work, busymod_work_func); 21 static void busymod_work_func(struct work_struct *work) argument 40 schedule_work(&work); 44 * exiting until the work function's entry message has printed. 50 * Serialize output: print all messages from the work 53 flush_work(&work); 62 flush_work(&work);
|
/linux-master/drivers/net/wireless/quantenna/qtnfmac/ |
H A D | event.h | 12 void qtnf_event_work_handler(struct work_struct *work);
|
/linux-master/fs/btrfs/ |
H A D | async-thread.c | 29 /* List head pointing to ordered work list */ 55 struct btrfs_fs_info * __pure btrfs_work_owner(const struct btrfs_work *work) argument 57 return work->wq->fs_info; 163 * Hook for threshold which will be called before executing the work, 214 struct btrfs_work *work; local 223 work = list_entry(list->next, struct btrfs_work, 225 if (!test_bit(WORK_DONE_BIT, &work->flags)) 231 * updates from ordinary work function. 237 * we leave the work item on the list as a barrier so 238 * that later work item 296 struct btrfs_work *work = container_of(normal_work, struct btrfs_work, local 331 btrfs_init_work(struct btrfs_work *work, btrfs_func_t func, btrfs_ordered_func_t ordered_func) argument 341 btrfs_queue_work(struct btrfs_workqueue *wq, struct btrfs_work *work) argument [all...] |
/linux-master/drivers/accessibility/speakup/ |
H A D | selection.c | 20 struct work_struct work; member in struct:speakup_selection_work 25 static void __speakup_set_selection(struct work_struct *work) argument 28 container_of(work, struct speakup_selection_work, work); 58 .work = __WORK_INITIALIZER(speakup_sel_work.work, 65 * cancelling selection work. getting kref first establishes the 87 schedule_work_on(WORK_CPU_UNBOUND, &speakup_sel_work.work); 96 cancel_work_sync(&speakup_sel_work.work); 97 /* setting to null so that if work fail 106 __speakup_paste_selection(struct work_struct *work) argument [all...] |