/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gpu_devfreq.c | 130 static void msm_devfreq_boost_work(struct kthread_work *work); 131 static void msm_devfreq_idle_work(struct kthread_work *work); 202 kthread_cancel_work_sync(&df->idle_work.work); 208 kthread_cancel_work_sync(&df->boost_work.work); 256 static void msm_devfreq_boost_work(struct kthread_work *work) argument 258 struct msm_gpu_devfreq *df = container_of(work, 259 struct msm_gpu_devfreq, boost_work.work); 315 * We could have become active again before the idle work had a 335 static void msm_devfreq_idle_work(struct kthread_work *work) argument 337 struct msm_gpu_devfreq *df = container_of(work, [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | hv_vhca.c | 70 static void mlx5_hv_vhca_invalidate_work(struct work_struct *work) argument 76 hwork = container_of(work, struct mlx5_hv_vhca_work, invalidate_work); 99 struct mlx5_hv_vhca_work *work; local 101 work = kzalloc(sizeof(*work), GFP_ATOMIC); 102 if (!work) 105 INIT_WORK(&work->invalidate_work, mlx5_hv_vhca_invalidate_work); 106 work->hv_vhca = hv_vhca; 107 work->block_mask = block_mask; 109 queue_work(hv_vhca->work_queue, &work [all...] |
/linux-master/drivers/media/platform/mediatek/mdp3/ |
H A D | mtk-mdp3-vpu.c | 35 if (!vpu->work) { 36 vpu->work = dma_alloc_wc(dev, vpu->work_size, 38 if (!vpu->work) 52 dma_free_wc(dev, vpu->work_size, vpu->work, vpu->work_addr); 53 vpu->work = NULL; 73 if (vpu->work && vpu->work_addr) 74 dma_free_wc(dev, vpu->work_size, vpu->work, vpu->work_addr); 224 "VPU param:%pK pa:%pad sz:%zx, work:%pK pa:%pad sz:%zx, config:%pK pa:%pad sz:%zx", 226 vpu->work, &vpu->work_addr, vpu->work_size, 274 memset(vpu->work, [all...] |
/linux-master/fs/smb/server/ |
H A D | smb2misc.c | 368 int ksmbd_smb2_check_message(struct ksmbd_work *work) argument 370 struct smb2_pdu *pdu = ksmbd_req_buf_next(work); 374 __u32 len = get_rfc1002_len(work->request_buf); 377 if ((u64)work->next_smb2_rcv_hdr_off + next_cmd > len) { 385 else if (work->next_smb2_rcv_hdr_off) 386 len -= work->next_smb2_rcv_hdr_off; 463 if ((work->conn->vals->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU) && 464 smb2_validate_credit_charge(work->conn, hdr)) 470 int smb2_negotiate_request(struct ksmbd_work *work) argument 472 return ksmbd_smb_negotiate_common(work, SMB2_NEGOTIATE_H [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_display.c | 51 * amdgpu_display_hotplug_work_func - work handler for display hotplug event 53 * @work: work struct pointer 55 * This is the hotplug event work handler (all ASICs). 56 * The work gets scheduled from the IRQ handler if there 62 * from the IRQ handler to a work handler because hotplug handler has to use 66 void amdgpu_display_hotplug_work_func(struct work_struct *work) argument 68 struct amdgpu_device *adev = container_of(work, struct amdgpu_device, 69 hotplug_work.work); 93 struct amdgpu_flip_work *work local 100 amdgpu_display_flip_handle_fence(struct amdgpu_flip_work *work, struct dma_fence **f) argument 122 struct amdgpu_flip_work *work = local 164 amdgpu_crtc->crtc_id, amdgpu_crtc, work); local 173 struct amdgpu_flip_work *work = local 200 struct amdgpu_flip_work *work; local [all...] |
/linux-master/drivers/gpu/drm/amd/display/amdgpu_dm/ |
H A D | amdgpu_dm_hdcp.c | 152 static void link_lock(struct hdcp_workqueue *work, bool lock) argument 156 for (i = 0; i < work->max_link; i++) { 158 mutex_lock(&work[i].mutex); 160 mutex_unlock(&work[i].mutex); 273 static void event_callback(struct work_struct *work) argument 277 hdcp_work = container_of(to_delayed_work(work), struct hdcp_workqueue, 292 static void event_property_update(struct work_struct *work) argument 294 struct hdcp_workqueue *hdcp_work = container_of(work, struct hdcp_workqueue, 363 static void event_property_validate(struct work_struct *work) argument 366 container_of(to_delayed_work(work), struc 415 event_watchdog_timer(struct work_struct *work) argument 436 event_cpirq(struct work_struct *work) argument 620 struct hdcp_workqueue *work; local 644 struct hdcp_workqueue *work; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ipsec.c | 62 container_of(_work, struct mlx5e_ipsec_dwork, dwork.work); 556 struct mlx5e_ipsec_work *work = local 557 container_of(_work, struct mlx5e_ipsec_work, work); 558 struct mlx5e_ipsec_sa_entry *sa_entry = work->sa_entry; 561 attrs = &((struct mlx5e_ipsec_sa_entry *)work->data)->attrs; 584 struct mlx5e_ipsec_work *work = local 585 container_of(_work, struct mlx5e_ipsec_work, work); 586 struct mlx5e_ipsec_sa_entry *sa_entry = work->sa_entry; 587 struct mlx5e_ipsec_netevent_data *data = work->data; 609 struct mlx5e_ipsec_work *work; local 973 struct mlx5e_ipsec_work *work = sa_entry->work; local [all...] |
/linux-master/security/keys/trusted-keys/ |
H A D | trusted_tpm2.c | 37 u8 *work = scratch, *work1; local 53 work = asn1_encode_oid(work, end_work, tpm2key_oid, 62 work = asn1_encode_tag(work, end_work, 0, bool, w - bool); 71 if (WARN(work - scratch + pub_len + priv_len + 14 > SCRATCH_SIZE, 75 work = asn1_encode_integer(work, end_work, options->keyhandle); 76 work = asn1_encode_octet_string(work, end_wor [all...] |
/linux-master/drivers/hv/ |
H A D | hv_util.c | 87 struct work_struct work; member in struct:hibernate_work_context 94 static void send_hibernate_uevent(struct work_struct *work) argument 99 ctx = container_of(work, struct hibernate_work_context, work); 110 INIT_WORK(&hibernate_context.work, send_hibernate_uevent); 188 struct work_struct *work = NULL; local 244 work = &shutdown_work; 250 work = &restart_work; 259 work = &hibernate_context.work; 343 hv_set_host_time(struct work_struct *work) argument [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | odp.c | 80 struct work_struct work; member in struct:mlx5_pagefault 190 static void free_implicit_child_mr_work(struct work_struct *work) argument 193 container_of(work, struct mlx5_ib_mr, odp_destroy.work); 221 /* Freeing a MR is a sleeping operation, so bounce to a work queue */ 222 INIT_WORK(&mr->odp_destroy.work, free_implicit_child_mr_work); 223 queue_work(system_unbound_wq, &mr->odp_destroy.work); 1343 * work-queue is being fenced. */ 1376 static void mlx5_ib_eqe_pf_action(struct work_struct *work) argument 1378 struct mlx5_pagefault *pfault = container_of(work, 1501 mlx5_ib_eq_pf_action(struct work_struct *work) argument 1643 struct work_struct work; member in struct:prefetch_mr_work 1653 destroy_prefetch_work(struct prefetch_mr_work *work) argument 1704 struct prefetch_mr_work *work = local 1724 init_prefetch_work(struct ib_pd *pd, enum ib_uverbs_advise_mr_advice advice, u32 pf_flags, struct prefetch_mr_work *work, struct ib_sge *sg_list, u32 num_sge) argument 1783 struct prefetch_mr_work *work; local [all...] |
/linux-master/sound/soc/codecs/ |
H A D | pcm1789.c | 31 struct work_struct work; member in struct:pcm1789_private 131 static void pcm1789_work_queue(struct work_struct *work) argument 133 struct pcm1789_private *priv = container_of(work, 135 work); 154 schedule_work(&priv->work); 254 INIT_WORK(&pcm1789->work, pcm1789_work_queue); 265 flush_work(&priv->work);
|
/linux-master/drivers/input/misc/ |
H A D | regulator-haptic.c | 25 struct work_struct work; member in struct:regulator_haptic 81 static void regulator_haptic_work(struct work_struct *work) argument 83 struct regulator_haptic *haptic = container_of(work, 84 struct regulator_haptic, work); 103 schedule_work(&haptic->work); 112 cancel_work_sync(&haptic->work); 157 INIT_WORK(&haptic->work, regulator_haptic_work);
|
H A D | drv2665.c | 50 * @work: Work item used to off load the enable/disable of the vibration 57 struct work_struct work; member in struct:drv2665_data 76 static void drv2665_worker(struct work_struct *work) argument 79 container_of(work, struct drv2665_data, work); 108 schedule_work(&haptics->work); 118 cancel_work_sync(&haptics->work); 196 INIT_WORK(&haptics->work, drv2665_worker);
|
H A D | pm8xxx-vibrator.c | 49 * @work: work structure to set the vibration parameters 59 struct work_struct work; member in struct:pm8xxx_vib 99 * @work: pointer to work_struct 101 static void pm8xxx_work_handler(struct work_struct *work) argument 103 struct pm8xxx_vib *vib = container_of(work, struct pm8xxx_vib, work); 139 cancel_work_sync(&vib->work); 161 schedule_work(&vib->work); 186 INIT_WORK(&vib->work, pm8xxx_work_handle [all...] |
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_gsc_uc.c | 16 static void gsc_work(struct work_struct *work) argument 18 struct intel_gsc_uc *gsc = container_of(work, typeof(*gsc), work); 128 INIT_WORK(&gsc->work, gsc_work); 166 * guaranteed to always work as long as the GPU itself is awake (which 258 flush_work(&gsc->work); 279 flush_work(&gsc->work); 315 queue_work(gsc->wq, &gsc->work);
|
/linux-master/mm/ |
H A D | page_reporting.c | 79 * Delay the start of work to allow a sizable queue to build. For 83 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); 307 static void page_reporting_process(struct work_struct *work) argument 309 struct delayed_work *d_work = to_delayed_work(work); 311 container_of(d_work, struct page_reporting_dev_info, work); 346 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); 379 /* initialize state and work structures */ 381 INIT_DELAYED_WORK(&prdev->work, &page_reporting_process); 411 /* Flush any existing work, and lock it out */ 412 cancel_delayed_work_sync(&prdev->work); [all...] |
/linux-master/kernel/irq/ |
H A D | irq_sim.c | 15 struct irq_work work; member in struct:irq_sim_work_ctx 81 irq_work_queue(&irq_ctx->work_ctx->work); 100 static void irq_sim_handle_irq(struct irq_work *work) argument 106 work_ctx = container_of(work, struct irq_sim_work_ctx, work); 184 work_ctx->work = IRQ_WORK_INIT_HARD(irq_sim_handle_irq); 201 irq_work_sync(&work_ctx->work);
|
/linux-master/drivers/net/wireless/realtek/rtw89/ |
H A D | coex.h | 212 void rtw89_btc_ntfy_eapol_packet_work(struct work_struct *work); 213 void rtw89_btc_ntfy_arp_packet_work(struct work_struct *work); 214 void rtw89_btc_ntfy_dhcp_packet_work(struct work_struct *work); 215 void rtw89_btc_ntfy_icmp_packet_work(struct work_struct *work); 226 void rtw89_coex_act1_work(struct work_struct *work); 227 void rtw89_coex_bt_devinfo_work(struct work_struct *work); 228 void rtw89_coex_rfk_chk_work(struct work_struct *work);
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_devcoredump.c | 46 * at least while the coredump device is alive. Dev_coredump has a delayed work 63 static void xe_devcoredump_deferred_snap_work(struct work_struct *work) argument 65 struct xe_devcoredump_snapshot *ss = container_of(work, typeof(*ss), work); 88 /* Ensure delayed work is captured before continuing */ 89 flush_work(&ss->work); 137 cancel_work_sync(&coredump->snapshot.work); 171 INIT_WORK(&ss->work, xe_devcoredump_deferred_snap_work); 200 queue_work(system_unbound_wq, &ss->work);
|
/linux-master/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | btcoex.c | 61 * @work: DHCP state machine work 83 struct work_struct work; member in struct:brcmf_btcoex_info 279 schedule_work(&bt_local->work); 283 * brcmf_btcoex_handler() - BT coex state machine work handler 284 * @work: work 286 static void brcmf_btcoex_handler(struct work_struct *work) argument 289 btci = container_of(work, struct brcmf_btcoex_info, work); [all...] |
/linux-master/drivers/isdn/mISDN/ |
H A D | timerdev.c | 32 u_int work; member in struct:mISDNtimerdev 57 dev->work = 0; 108 while (list_empty(list) && (dev->work == 0)) { 112 wait_event_interruptible(dev->wait, (dev->work || 118 if (dev->work) 119 dev->work = 0; 146 if (dev->work || !list_empty(&dev->expired)) 149 printk(KERN_DEBUG "%s work(%d) empty(%d)\n", __func__, 150 dev->work, list_empty(&dev->expired)); 175 dev->work [all...] |
/linux-master/sound/soc/ |
H A D | soc-jack.c | 243 queue_delayed_work(system_power_efficient_wq, &gpio->work, 249 /* gpio work */ 250 static void gpio_work(struct work_struct *work) argument 254 gpio = container_of(work, struct snd_soc_jack_gpio, work.work); 270 * concurrently with work triggered by the interrupt handler. 272 queue_delayed_work(system_power_efficient_wq, &gpio->work, 0); 288 cancel_delayed_work_sync(&gpios[i].work); 365 INIT_DELAYED_WORK(&gpios[i].work, gpio_wor [all...] |
/linux-master/drivers/input/touchscreen/ |
H A D | stmpe-ts.c | 53 * @work: a work item used to scan the device 75 struct delayed_work work; member in struct:stmpe_touch 98 static void stmpe_work(struct work_struct *work) argument 104 container_of(work, struct stmpe_touch, work.work); 139 cancel_delayed_work_sync(&ts->work); 169 schedule_delayed_work(&ts->work, msecs_to_jiffies(50)); 253 cancel_delayed_work_sync(&ts->work); [all...] |
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_power.c | 92 cancel_delayed_work_sync(&pvr_dev->watchdog.work); 122 queue_delayed_work(pvr_dev->sched_wq, &pvr_dev->watchdog.work, 183 pvr_watchdog_worker(struct work_struct *work) argument 185 struct pvr_device *pvr_dev = container_of(work, struct pvr_device, 186 watchdog.work.work); 212 queue_delayed_work(pvr_dev->sched_wq, &pvr_dev->watchdog.work, 228 INIT_DELAYED_WORK(&pvr_dev->watchdog.work, pvr_watchdog_worker); 432 cancel_delayed_work_sync(&pvr_dev->watchdog.work);
|
/linux-master/fs/xfs/ |
H A D | xfs_mru_cache.c | 14 * queueing up work in the form of timed callbacks is also included. 44 * The current implementation uses work queue callbacks to carefully time the 101 struct delayed_work work; /* Workqueue data for reaping. */ member in struct:xfs_mru_cache 102 unsigned int queued; /* work has been queued */ 123 * zero if there's no need to schedule more work because the lists are empty. 199 * zero and start the work queue timer if necessary. Otherwise, set grp 206 queue_delayed_work(xfs_mru_reap_wq, &mru->work, 266 struct work_struct *work) 269 container_of(work, struct xfs_mru_cache, work 265 _xfs_mru_cache_reap( struct work_struct *work) argument [all...] |