/linux-master/drivers/gpu/drm/ |
H A D | drm_suballoc.c | 75 init_waitqueue_head(&sa_manager->wq); 218 spin_lock(&sa_manager->wq.lock); 220 spin_unlock(&sa_manager->wq.lock); 338 spin_lock(&sa_manager->wq.lock); 348 spin_unlock(&sa_manager->wq.lock); 362 spin_unlock(&sa_manager->wq.lock); 370 spin_lock(&sa_manager->wq.lock); 374 (sa_manager->wq, 377 spin_unlock(&sa_manager->wq.lock); 378 wait_event(sa_manager->wq, [all...] |
/linux-master/include/linux/ |
H A D | wait.h | 172 struct wait_queue_entry *wq; local 174 list_for_each_entry(wq, &wq_head->head, entry) { 175 if (!(wq->flags & WQ_FLAG_PRIORITY)) 177 head = &wq->entry; 592 * @wq: the waitqueue to wait on 598 * The @condition is checked each time the waitqueue @wq is woken up. 606 #define wait_event_interruptible_hrtimeout(wq, condition, timeout) \ 611 __ret = __wait_event_hrtimeout(wq, condition, timeout, \ 616 #define __wait_event_interruptible_exclusive(wq, condition) \ 617 ___wait_event(wq, conditio [all...] |
H A D | closure.h | 146 struct workqueue_struct *wq; member in struct:closure::__anon162::__anon163 236 struct workqueue_struct *wq) 240 cl->wq = wq; 245 struct workqueue_struct *wq = cl->wq; local 253 if (wq) { 255 BUG_ON(!queue_work(wq, &cl->work)); 322 * of @wq (or, if @wq i 235 set_closure_fn(struct closure *cl, closure_fn *fn, struct workqueue_struct *wq) argument 385 closure_call(struct closure *cl, closure_fn fn, struct workqueue_struct *wq, struct closure *parent) argument [all...] |
H A D | sync_file.h | 27 * @wq: wait queue for fence signaling 49 wait_queue_head_t wq; member in struct:sync_file
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.h | 42 #include "wq.h" 56 struct mlx5_cqwq wq; member in struct:mlx5_fpga_conn::__anon1227 66 struct mlx5_wq_qp wq; member in struct:mlx5_fpga_conn::__anon1228
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | mmu_rb.h | 49 struct workqueue_struct *wq; member in struct:mmu_rb_handler 55 struct workqueue_struct *wq,
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_utils.h | 64 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_static_params_wqe))) 68 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_progress_params_wqe))) 72 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_get_tls_progress_params_wqe))) 76 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_dump_wqe)))
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.c | 687 .wq = c->priv->wq, 965 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local 986 MLX5_SET(wq, wq, log_wqe_num_of_strides, 988 MLX5_SET(wq, wq, log_wqe_stride_size, 990 MLX5_SET(wq, wq, log_wq_s 1036 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local 1061 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1074 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1262 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1276 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1294 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local [all...] |
/linux-master/tools/workqueue/ |
H A D | wq_monitor.py | 77 def __init__(self, wq): 78 self.name = wq.name.string_().decode() 79 self.unbound = wq.flags & WQ_UNBOUND != 0 80 self.mem_reclaim = wq.flags & WQ_MEM_RECLAIM != 0 82 for pwq in list_for_each_entry('struct pool_workqueue', wq.pwqs.address_of_(), 'pwqs_node'): 161 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'): 162 stats = WqStats(wq)
|
/linux-master/net/core/ |
H A D | stream.c | 33 struct socket_wq *wq; local 39 wq = rcu_dereference(sk->sk_wq); 40 if (skwq_has_sleeper(wq)) 41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | 43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) 44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT);
|
/linux-master/net/sunrpc/xprtrdma/ |
H A D | svc_rdma.c | 266 struct workqueue_struct *wq = svcrdma_wq; local 269 destroy_workqueue(wq); 277 struct workqueue_struct *wq; local 280 wq = alloc_workqueue("svcrdma", WQ_UNBOUND, 0); 281 if (!wq) 286 destroy_workqueue(wq); 290 svcrdma_wq = wq;
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_debugfs.c | 25 struct hinic_wq *wq = sq->wq; local 31 return atomic_read(&wq->prod_idx) & wq->mask; 33 return atomic_read(&wq->cons_idx) & wq->mask; 35 return be16_to_cpu(*(__be16 *)(sq->hw_ci_addr)) & wq->mask; 55 struct hinic_wq *wq = rq->wq; local 61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq [all...] |
/linux-master/drivers/hwmon/ |
H A D | nzxt-smart2.c | 189 * wq is used to wait for *_received flags to become true. 191 * wq.lock held. 193 wait_queue_head_t wq; member in struct:drvdata 232 spin_lock(&drvdata->wq.lock); 238 wake_up_all_locked(&drvdata->wq); 239 spin_unlock(&drvdata->wq.lock); 250 spin_lock(&drvdata->wq.lock); 259 spin_unlock(&drvdata->wq.lock); 290 wake_up_all_locked(&drvdata->wq); 302 wake_up_all_locked(&drvdata->wq); [all...] |
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_hw_queues.c | 16 struct hw_wq *wq = NULL; local 65 wq = efct_hw_new_wq(cq, hw->num_qentries[SLI4_QTYPE_WQ]); 66 if (!wq) { 278 struct hw_wq *wq = kzalloc(sizeof(*wq), GFP_KERNEL); local 280 if (!wq) 283 wq->hw = cq->eq->hw; 284 wq->cq = cq; 285 wq->type = SLI4_QTYPE_WQ; 286 wq 444 efct_hw_del_wq(struct hw_wq *wq) argument [all...] |
/linux-master/drivers/dma/idxd/ |
H A D | init.c | 150 struct idxd_wq *wq; local 166 wq = kzalloc_node(sizeof(*wq), GFP_KERNEL, dev_to_node(dev)); 167 if (!wq) { 172 idxd_dev_set_type(&wq->idxd_dev, IDXD_DEV_WQ); 173 conf_dev = wq_confdev(wq); 174 wq->id = i; 175 wq->idxd = idxd; 176 device_initialize(wq_confdev(wq)); 180 rc = dev_set_name(conf_dev, "wq 792 struct idxd_wq *wq; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_sched.h | 102 struct workqueue_struct *wq; member in struct:nouveau_sched 110 struct wait_queue_head wq; member in struct:nouveau_sched::__anon25 115 struct workqueue_struct *wq, u32 credit_limit);
|
H A D | nouveau_sched.c | 127 wake_up(&sched->job.wq); 403 struct workqueue_struct *wq, u32 credit_limit) 410 if (!wq) { 411 wq = alloc_workqueue("nouveau_sched_wq_%d", 0, WQ_MAX_ACTIVE, 413 if (!wq) 416 sched->wq = wq; 419 ret = drm_sched_init(drm_sched, &nouveau_sched_ops, wq, 444 init_waitqueue_head(&sched->job.wq); 451 if (sched->wq) 402 nouveau_sched_init(struct nouveau_sched *sched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument 457 nouveau_sched_create(struct nouveau_sched **psched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument [all...] |
/linux-master/include/drm/ |
H A D | drm_debugfs_crc.h | 51 * @wq: workqueue used to synchronize reading and writing 60 wait_queue_head_t wq; member in struct:drm_crtc_crc
|
H A D | drm_flip_work.h | 73 struct workqueue_struct *wq);
|
/linux-master/include/uapi/rdma/ |
H A D | rvt-abi.h | 54 * Note that the wq array elements are variable size so you can't 64 struct rvt_rwqe wq[]; member in struct:rvt_rwq
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
H A D | mp.h | 24 struct workqueue_struct *wq; member in struct:lag_mp
|
/linux-master/drivers/usb/chipidea/ |
H A D | otg.h | 21 if (queue_work(ci->wq, &ci->work) == false)
|
/linux-master/drivers/md/bcache/ |
H A D | request.h | 9 struct workqueue_struct *wq; member in struct:data_insert_op
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_gsc_types.h | 37 /** @wq: workqueue to handle jobs for delayed load and proxy handling */ 38 struct workqueue_struct *wq; member in struct:xe_gsc
|
/linux-master/drivers/media/platform/mediatek/vcodec/decoder/ |
H A D | vdec_vpu_if.h | 25 * @wq : wait queue to wait VPU message ack 41 wait_queue_head_t wq; member in struct:vdec_vpu_inst
|