/linux-master/tools/workqueue/ |
H A D | wq_dump.py | 81 def wq_type_str(wq): 82 if wq.flags & WQ_BH: 84 elif wq.flags & WQ_UNBOUND: 85 if wq.flags & WQ_ORDERED: 88 if wq.unbound_attrs.affn_strict: 181 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'): 182 print(f'{wq.name.string_().decode():{WQ_NAME_LEN}} {wq_type_str(wq):10}', end='') 185 pool_id = per_cpu_ptr(wq.cpu_pwq, cpu)[0].pool.id.value_() 189 if wq [all...] |
H A D | wq_monitor.py | 70 def __init__(self, wq): 71 self.name = wq.name.string_().decode() 72 self.unbound = wq.flags & WQ_UNBOUND != 0 73 self.mem_reclaim = wq.flags & WQ_MEM_RECLAIM != 0 75 for pwq in list_for_each_entry('struct pool_workqueue', wq.pwqs.address_of_(), 'pwqs_node'): 154 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'): 155 stats = WqStats(wq)
|
/linux-master/fs/btrfs/ |
H A D | misc.h | 21 static inline void cond_wake_up(struct wait_queue_head *wq) argument 27 if (wq_has_sleeper(wq)) 28 wake_up(wq); 31 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) argument 39 if (waitqueue_active(wq)) 40 wake_up(wq);
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | srq_cmd.c | 27 static void set_wq(void *wq, struct mlx5_srq_attr *in) argument 29 MLX5_SET(wq, wq, wq_signature, !!(in->flags 31 MLX5_SET(wq, wq, log_wq_pg_sz, in->log_page_size); 32 MLX5_SET(wq, wq, log_wq_stride, in->wqe_shift + 4); 33 MLX5_SET(wq, wq, log_wq_sz, in->log_size); 34 MLX5_SET(wq, w 55 get_wq(void *wq, struct mlx5_srq_attr *in) argument 334 void *wq; local 402 void *wq; local 481 void *wq; local [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_suballoc.c | 75 init_waitqueue_head(&sa_manager->wq); 218 spin_lock(&sa_manager->wq.lock); 220 spin_unlock(&sa_manager->wq.lock); 338 spin_lock(&sa_manager->wq.lock); 348 spin_unlock(&sa_manager->wq.lock); 362 spin_unlock(&sa_manager->wq.lock); 370 spin_lock(&sa_manager->wq.lock); 374 (sa_manager->wq, 377 spin_unlock(&sa_manager->wq.lock); 378 wait_event(sa_manager->wq, [all...] |
/linux-master/include/linux/ |
H A D | wait.h | 172 struct wait_queue_entry *wq; local 174 list_for_each_entry(wq, &wq_head->head, entry) { 175 if (!(wq->flags & WQ_FLAG_PRIORITY)) 177 head = &wq->entry; 592 * @wq: the waitqueue to wait on 598 * The @condition is checked each time the waitqueue @wq is woken up. 606 #define wait_event_interruptible_hrtimeout(wq, condition, timeout) \ 611 __ret = __wait_event_hrtimeout(wq, condition, timeout, \ 616 #define __wait_event_interruptible_exclusive(wq, condition) \ 617 ___wait_event(wq, conditio [all...] |
H A D | closure.h | 146 struct workqueue_struct *wq; member in struct:closure::__anon147::__anon148 248 struct workqueue_struct *wq) 252 cl->wq = wq; 257 struct workqueue_struct *wq = cl->wq; local 265 if (wq) { 267 BUG_ON(!queue_work(wq, &cl->work)); 334 * of @wq (or, if @wq i 247 set_closure_fn(struct closure *cl, closure_fn *fn, struct workqueue_struct *wq) argument 397 closure_call(struct closure *cl, closure_fn fn, struct workqueue_struct *wq, struct closure *parent) argument [all...] |
H A D | sync_file.h | 27 * @wq: wait queue for fence signaling 49 wait_queue_head_t wq; member in struct:sync_file
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | conn.h | 42 #include "wq.h" 56 struct mlx5_cqwq wq; member in struct:mlx5_fpga_conn::__anon1227 66 struct mlx5_wq_qp wq; member in struct:mlx5_fpga_conn::__anon1228
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | mmu_rb.h | 49 struct workqueue_struct *wq; member in struct:mmu_rb_handler 55 struct workqueue_struct *wq,
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_utils.h | 64 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_static_params_wqe))) 68 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_progress_params_wqe))) 72 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_get_tls_progress_params_wqe))) 76 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_dump_wqe)))
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | params.c | 617 .wq = c->priv->wq, 895 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local 916 MLX5_SET(wq, wq, log_wqe_num_of_strides, 918 MLX5_SET(wq, wq, log_wqe_stride_size, 920 MLX5_SET(wq, wq, log_wq_s 966 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local 991 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1004 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1192 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1206 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local 1224 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local [all...] |
/linux-master/net/core/ |
H A D | stream.c | 33 struct socket_wq *wq; local 39 wq = rcu_dereference(sk->sk_wq); 40 if (skwq_has_sleeper(wq)) 41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | 43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) 44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT);
|
/linux-master/net/sunrpc/xprtrdma/ |
H A D | svc_rdma.c | 265 struct workqueue_struct *wq = svcrdma_wq; local 268 destroy_workqueue(wq); 276 struct workqueue_struct *wq; local 279 wq = alloc_workqueue("svcrdma", WQ_UNBOUND, 0); 280 if (!wq) 285 destroy_workqueue(wq); 289 svcrdma_wq = wq;
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_debugfs.c | 25 struct hinic_wq *wq = sq->wq; local 31 return atomic_read(&wq->prod_idx) & wq->mask; 33 return atomic_read(&wq->cons_idx) & wq->mask; 35 return be16_to_cpu(*(__be16 *)(sq->hw_ci_addr)) & wq->mask; 55 struct hinic_wq *wq = rq->wq; local 61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq [all...] |
/linux-master/drivers/hwmon/ |
H A D | nzxt-smart2.c | 189 * wq is used to wait for *_received flags to become true. 191 * wq.lock held. 193 wait_queue_head_t wq; member in struct:drvdata 232 spin_lock(&drvdata->wq.lock); 238 wake_up_all_locked(&drvdata->wq); 239 spin_unlock(&drvdata->wq.lock); 250 spin_lock(&drvdata->wq.lock); 259 spin_unlock(&drvdata->wq.lock); 290 wake_up_all_locked(&drvdata->wq); 302 wake_up_all_locked(&drvdata->wq); [all...] |
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_hw_queues.c | 16 struct hw_wq *wq = NULL; local 65 wq = efct_hw_new_wq(cq, hw->num_qentries[SLI4_QTYPE_WQ]); 66 if (!wq) { 278 struct hw_wq *wq = kzalloc(sizeof(*wq), GFP_KERNEL); local 280 if (!wq) 283 wq->hw = cq->eq->hw; 284 wq->cq = cq; 285 wq->type = SLI4_QTYPE_WQ; 286 wq 444 efct_hw_del_wq(struct hw_wq *wq) argument [all...] |
/linux-master/drivers/dma/idxd/ |
H A D | init.c | 152 struct idxd_wq *wq; local 168 wq = kzalloc_node(sizeof(*wq), GFP_KERNEL, dev_to_node(dev)); 169 if (!wq) { 174 idxd_dev_set_type(&wq->idxd_dev, IDXD_DEV_WQ); 175 conf_dev = wq_confdev(wq); 176 wq->id = i; 177 wq->idxd = idxd; 178 device_initialize(wq_confdev(wq)); 182 rc = dev_set_name(conf_dev, "wq 796 struct idxd_wq *wq; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_sched.h | 102 struct workqueue_struct *wq; member in struct:nouveau_sched 110 struct wait_queue_head wq; member in struct:nouveau_sched::__anon25 115 struct workqueue_struct *wq, u32 credit_limit);
|
H A D | nouveau_sched.c | 127 wake_up(&sched->job.wq); 403 struct workqueue_struct *wq, u32 credit_limit) 410 if (!wq) { 411 wq = alloc_workqueue("nouveau_sched_wq_%d", 0, WQ_MAX_ACTIVE, 413 if (!wq) 416 sched->wq = wq; 419 ret = drm_sched_init(drm_sched, &nouveau_sched_ops, wq, 444 init_waitqueue_head(&sched->job.wq); 451 if (sched->wq) 402 nouveau_sched_init(struct nouveau_sched *sched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument 457 nouveau_sched_create(struct nouveau_sched **psched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument [all...] |
/linux-master/include/uapi/rdma/ |
H A D | rvt-abi.h | 54 * Note that the wq array elements are variable size so you can't 64 struct rvt_rwqe wq[]; member in struct:rvt_rwq
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
H A D | mp.h | 24 struct workqueue_struct *wq; member in struct:lag_mp
|
/linux-master/drivers/usb/chipidea/ |
H A D | otg.h | 21 if (queue_work(ci->wq, &ci->work) == false)
|
/linux-master/drivers/md/bcache/ |
H A D | request.h | 9 struct workqueue_struct *wq; member in struct:data_insert_op
|
/linux-master/include/drm/ |
H A D | drm_flip_work.h | 73 struct workqueue_struct *wq);
|