Searched refs:wq (Results 51 - 75 of 700) sorted by relevance

1234567891011>>

/linux-master/drivers/gpu/drm/
H A Ddrm_suballoc.c75 init_waitqueue_head(&sa_manager->wq);
218 spin_lock(&sa_manager->wq.lock);
220 spin_unlock(&sa_manager->wq.lock);
338 spin_lock(&sa_manager->wq.lock);
348 spin_unlock(&sa_manager->wq.lock);
362 spin_unlock(&sa_manager->wq.lock);
370 spin_lock(&sa_manager->wq.lock);
374 (sa_manager->wq,
377 spin_unlock(&sa_manager->wq.lock);
378 wait_event(sa_manager->wq,
[all...]
/linux-master/include/linux/
H A Dwait.h172 struct wait_queue_entry *wq; local
174 list_for_each_entry(wq, &wq_head->head, entry) {
175 if (!(wq->flags & WQ_FLAG_PRIORITY))
177 head = &wq->entry;
592 * @wq: the waitqueue to wait on
598 * The @condition is checked each time the waitqueue @wq is woken up.
606 #define wait_event_interruptible_hrtimeout(wq, condition, timeout) \
611 __ret = __wait_event_hrtimeout(wq, condition, timeout, \
616 #define __wait_event_interruptible_exclusive(wq, condition) \
617 ___wait_event(wq, conditio
[all...]
H A Dclosure.h146 struct workqueue_struct *wq; member in struct:closure::__anon162::__anon163
236 struct workqueue_struct *wq)
240 cl->wq = wq;
245 struct workqueue_struct *wq = cl->wq; local
253 if (wq) {
255 BUG_ON(!queue_work(wq, &cl->work));
322 * of @wq (or, if @wq i
235 set_closure_fn(struct closure *cl, closure_fn *fn, struct workqueue_struct *wq) argument
385 closure_call(struct closure *cl, closure_fn fn, struct workqueue_struct *wq, struct closure *parent) argument
[all...]
H A Dsync_file.h27 * @wq: wait queue for fence signaling
49 wait_queue_head_t wq; member in struct:sync_file
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.h42 #include "wq.h"
56 struct mlx5_cqwq wq; member in struct:mlx5_fpga_conn::__anon1227
66 struct mlx5_wq_qp wq; member in struct:mlx5_fpga_conn::__anon1228
/linux-master/drivers/infiniband/hw/hfi1/
H A Dmmu_rb.h49 struct workqueue_struct *wq; member in struct:mmu_rb_handler
55 struct workqueue_struct *wq,
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dktls_utils.h64 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_static_params_wqe)))
68 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_set_tls_progress_params_wqe)))
72 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_get_tls_progress_params_wqe)))
76 mlx5e_fetch_wqe(&(sq)->wq, pi, sizeof(struct mlx5e_dump_wqe)))
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dparams.c687 .wq = c->priv->wq,
965 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local
986 MLX5_SET(wq, wq, log_wqe_num_of_strides,
988 MLX5_SET(wq, wq, log_wqe_stride_size,
990 MLX5_SET(wq, wq, log_wq_s
1036 void *wq = MLX5_ADDR_OF(rqc, rqc, wq); local
1061 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local
1074 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local
1262 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local
1276 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local
1294 void *wq = MLX5_ADDR_OF(sqc, sqc, wq); local
[all...]
/linux-master/tools/workqueue/
H A Dwq_monitor.py77 def __init__(self, wq):
78 self.name = wq.name.string_().decode()
79 self.unbound = wq.flags & WQ_UNBOUND != 0
80 self.mem_reclaim = wq.flags & WQ_MEM_RECLAIM != 0
82 for pwq in list_for_each_entry('struct pool_workqueue', wq.pwqs.address_of_(), 'pwqs_node'):
161 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'):
162 stats = WqStats(wq)
/linux-master/net/core/
H A Dstream.c33 struct socket_wq *wq; local
39 wq = rcu_dereference(sk->sk_wq);
40 if (skwq_has_sleeper(wq))
41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT |
43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN))
44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT);
/linux-master/net/sunrpc/xprtrdma/
H A Dsvc_rdma.c266 struct workqueue_struct *wq = svcrdma_wq; local
269 destroy_workqueue(wq);
277 struct workqueue_struct *wq; local
280 wq = alloc_workqueue("svcrdma", WQ_UNBOUND, 0);
281 if (!wq)
286 destroy_workqueue(wq);
290 svcrdma_wq = wq;
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_debugfs.c25 struct hinic_wq *wq = sq->wq; local
31 return atomic_read(&wq->prod_idx) & wq->mask;
33 return atomic_read(&wq->cons_idx) & wq->mask;
35 return be16_to_cpu(*(__be16 *)(sq->hw_ci_addr)) & wq->mask;
55 struct hinic_wq *wq = rq->wq; local
61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq
[all...]
/linux-master/drivers/hwmon/
H A Dnzxt-smart2.c189 * wq is used to wait for *_received flags to become true.
191 * wq.lock held.
193 wait_queue_head_t wq; member in struct:drvdata
232 spin_lock(&drvdata->wq.lock);
238 wake_up_all_locked(&drvdata->wq);
239 spin_unlock(&drvdata->wq.lock);
250 spin_lock(&drvdata->wq.lock);
259 spin_unlock(&drvdata->wq.lock);
290 wake_up_all_locked(&drvdata->wq);
302 wake_up_all_locked(&drvdata->wq);
[all...]
/linux-master/drivers/scsi/elx/efct/
H A Defct_hw_queues.c16 struct hw_wq *wq = NULL; local
65 wq = efct_hw_new_wq(cq, hw->num_qentries[SLI4_QTYPE_WQ]);
66 if (!wq) {
278 struct hw_wq *wq = kzalloc(sizeof(*wq), GFP_KERNEL); local
280 if (!wq)
283 wq->hw = cq->eq->hw;
284 wq->cq = cq;
285 wq->type = SLI4_QTYPE_WQ;
286 wq
444 efct_hw_del_wq(struct hw_wq *wq) argument
[all...]
/linux-master/drivers/dma/idxd/
H A Dinit.c150 struct idxd_wq *wq; local
166 wq = kzalloc_node(sizeof(*wq), GFP_KERNEL, dev_to_node(dev));
167 if (!wq) {
172 idxd_dev_set_type(&wq->idxd_dev, IDXD_DEV_WQ);
173 conf_dev = wq_confdev(wq);
174 wq->id = i;
175 wq->idxd = idxd;
176 device_initialize(wq_confdev(wq));
180 rc = dev_set_name(conf_dev, "wq
792 struct idxd_wq *wq; local
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_sched.h102 struct workqueue_struct *wq; member in struct:nouveau_sched
110 struct wait_queue_head wq; member in struct:nouveau_sched::__anon25
115 struct workqueue_struct *wq, u32 credit_limit);
H A Dnouveau_sched.c127 wake_up(&sched->job.wq);
403 struct workqueue_struct *wq, u32 credit_limit)
410 if (!wq) {
411 wq = alloc_workqueue("nouveau_sched_wq_%d", 0, WQ_MAX_ACTIVE,
413 if (!wq)
416 sched->wq = wq;
419 ret = drm_sched_init(drm_sched, &nouveau_sched_ops, wq,
444 init_waitqueue_head(&sched->job.wq);
451 if (sched->wq)
402 nouveau_sched_init(struct nouveau_sched *sched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument
457 nouveau_sched_create(struct nouveau_sched **psched, struct nouveau_drm *drm, struct workqueue_struct *wq, u32 credit_limit) argument
[all...]
/linux-master/include/drm/
H A Ddrm_debugfs_crc.h51 * @wq: workqueue used to synchronize reading and writing
60 wait_queue_head_t wq; member in struct:drm_crtc_crc
H A Ddrm_flip_work.h73 struct workqueue_struct *wq);
/linux-master/include/uapi/rdma/
H A Drvt-abi.h54 * Note that the wq array elements are variable size so you can't
64 struct rvt_rwqe wq[]; member in struct:rvt_rwq
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lag/
H A Dmp.h24 struct workqueue_struct *wq; member in struct:lag_mp
/linux-master/drivers/usb/chipidea/
H A Dotg.h21 if (queue_work(ci->wq, &ci->work) == false)
/linux-master/drivers/md/bcache/
H A Drequest.h9 struct workqueue_struct *wq; member in struct:data_insert_op
/linux-master/drivers/gpu/drm/xe/
H A Dxe_gsc_types.h37 /** @wq: workqueue to handle jobs for delayed load and proxy handling */
38 struct workqueue_struct *wq; member in struct:xe_gsc
/linux-master/drivers/media/platform/mediatek/vcodec/decoder/
H A Dvdec_vpu_if.h25 * @wq : wait queue to wait VPU message ack
41 wait_queue_head_t wq; member in struct:vdec_vpu_inst

Completed in 264 milliseconds

1234567891011>>