Searched refs:wq (Results 76 - 100 of 700) sorted by relevance

1234567891011>>

/linux-master/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h422 struct lpfc_queue *wq; local
426 wq = phba->sli4_hba.hdwq[wqidx].io_wq;
429 wq = phba->sli4_hba.mbx_wq;
432 wq = phba->sli4_hba.els_wq;
435 wq = phba->sli4_hba.nvmels_wq;
442 qtypestr, wqidx, wq->queue_id);
445 qtypestr, wq->queue_id);
447 lpfc_debug_dump_q(wq);
462 struct lpfc_queue *wq, *cq, *eq; local
466 /* io wq an
[all...]
/linux-master/drivers/char/tpm/
H A Dtpm_ibmvtpm.h29 wait_queue_head_t wq; member in struct:ibmvtpm_crq_queue
41 wait_queue_head_t wq; member in struct:ibmvtpm_dev
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_qp.h54 #define HINIC_MIN_TX_WQE_SIZE(wq) \
55 ALIGN(HINIC_SQ_WQE_SIZE(1), (wq)->wqebb_size)
58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size)
82 struct hinic_wq *wq; member in struct:hinic_sq
101 struct hinic_wq *wq; member in struct:hinic_rq
139 struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr,
145 struct hinic_wq *wq, struct msix_entry *entry);
H A Dhinic_hw_cmdq.c343 struct hinic_wq *wq = cmdq->wq; local
350 /* WQE_SIZE = WQEBB_SIZE, we will get the wq element and not shadow*/
351 hw_wqe = hinic_get_wqe(wq, WQE_LCMD_SIZE, &curr_prod_idx);
361 num_wqebbs = ALIGN(WQE_LCMD_SIZE, wq->wqebb_size) / wq->wqebb_size;
363 if (next_prod_idx >= wq->q_depth) {
365 next_prod_idx -= wq->q_depth;
380 /* CMDQ WQE is not shadow, therefore wqe will be written to wq */
422 struct hinic_wq *wq local
699 struct hinic_wq *wq = cmdq->wq; local
738 init_cmdq(struct hinic_cmdq *cmdq, struct hinic_wq *wq, enum hinic_cmdq_type q_type, void __iomem *db_area) argument
[all...]
/linux-master/drivers/power/supply/
H A Dipaq_micro_battery.c40 struct workqueue_struct *wq; member in struct:micro_battery
88 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
235 mb->wq = alloc_workqueue("ipaq-battery-wq", WQ_MEM_RECLAIM, 0);
236 if (!mb->wq)
241 queue_delayed_work(mb->wq, &mb->update, 1);
264 destroy_workqueue(mb->wq);
276 destroy_workqueue(mb->wq);
291 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
/linux-master/drivers/gpu/drm/xe/
H A Dxe_guc_submit_types.h60 /** @wq: Workqueue info */
61 u32 wq[WQ_SIZE / sizeof(u32)]; member in struct:guc_submit_parallel_scratch
145 /** @wq: Workqueue Items */
146 u32 wq[WQ_SIZE / sizeof(u32)]; member in struct:xe_guc_submit_exec_queue_snapshot::__anon626
/linux-master/tools/testing/selftests/net/
H A Dioam6.sh308 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.ioam6_id=${ALPHA[0]}
309 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.ioam6_id_wide=${ALPHA[1]}
310 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.conf.veth0.ioam6_id=${ALPHA[4]}
311 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.conf.veth0.ioam6_id_wide=${ALPHA[5]}
316 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.all.forwarding=1
317 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.ioam6_id=${BETA[0]}
318 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.ioam6_id_wide=${BETA[1]}
319 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.veth0.ioam6_enabled=1
320 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.veth0.ioam6_id=${BETA[2]}
321 ip netns exec $ioam_node_beta sysctl -wq ne
[all...]
H A Dsrv6_end_dt46_l3vpn_test.sh264 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.accept_dad=0
265 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.default.accept_dad=0
271 ip netns exec ${nsname} sysctl -wq net.ipv4.ip_forward=1
272 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.forwarding=1
285 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.all.accept_dad=0
286 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.default.accept_dad=0
290 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.all.rp_filter=0
291 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.default.rp_filter=0
305 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.all.accept_dad=0
306 ip netns exec ${rtname} sysctl -wq ne
[all...]
/linux-master/drivers/infiniband/hw/mana/
H A Dqp.c110 struct mana_ib_wq *wq; local
192 wq = container_of(ibwq, struct mana_ib_wq, ibwq);
197 wq_spec.gdma_region = wq->gdma_region;
198 wq_spec.queue_size = wq->wq_buf_size;
207 &wq_spec, &cq_spec, &wq->rx_object);
215 wq->gdma_region = GDMA_INVALID_DMA_REGION;
218 wq->id = wq_spec.queue_index;
222 "ret %d rx_object 0x%llx wq id %llu cq id %llu\n",
223 ret, wq->rx_object, wq
477 struct mana_ib_wq *wq; local
[all...]
/linux-master/drivers/hid/
H A Dhid-elo.c34 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct
182 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL);
258 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL);
296 wq = create_singlethread_workqueue("elousb");
297 if (!wq)
302 destroy_workqueue(wq);
311 destroy_workqueue(wq);
/linux-master/drivers/i2c/busses/
H A Di2c-taos-evm.c30 static DECLARE_WAIT_QUEUE_HEAD(wq);
104 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
161 wake_up_interruptible(&wq);
166 wake_up_interruptible(&wq);
173 wake_up_interruptible(&wq);
226 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
248 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
/linux-master/block/
H A Dblk-rq-qos.c198 struct wait_queue_entry wq; member in struct:rq_qos_wait_data
211 wq);
248 .wq = {
250 .entry = LIST_HEAD_INIT(data.wq.entry),
263 has_sleeper = !prepare_to_wait_exclusive(&rqw->wait, &data.wq,
270 finish_wait(&rqw->wait, &data.wq);
286 finish_wait(&rqw->wait, &data.wq);
/linux-master/drivers/platform/x86/
H A Dasus-wireless.c27 struct workqueue_struct *wq; member in struct:asus_wireless_data
108 queue_work(data->wq, &data->led_work);
157 data->wq = create_singlethread_workqueue("asus_wireless_workqueue");
158 if (!data->wq)
169 destroy_workqueue(data->wq);
178 if (data->wq) {
180 destroy_workqueue(data->wq);
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_rx.c81 static inline void mlx5e_read_cqe_slot(struct mlx5_cqwq *wq, argument
84 u32 ci = mlx5_cqwq_ctr2ix(wq, cqcc);
86 memcpy(data, mlx5_cqwq_get_wqe(wq, ci), sizeof(struct mlx5_cqe64));
105 mlx5_wq_cyc_ctr2ix(&rq->wqe.wq, be16_to_cpu(title->wqe_counter) + 1);
109 struct mlx5_cqwq *wq,
115 mlx5e_read_cqe_slot(wq, cqcc, title);
121 static inline void mlx5e_read_mini_arr_slot(struct mlx5_cqwq *wq, argument
125 mlx5e_read_cqe_slot(wq, cqcc, cqd->mini_arr);
129 static inline void mlx5e_cqes_update_owner(struct mlx5_cqwq *wq, int n) argument
131 u32 cqcc = wq
108 mlx5e_read_title_slot(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument
153 mlx5e_decompress_cqe(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument
183 mlx5e_decompress_cqe_no_hash(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument
194 mlx5e_decompress_enhanced_cqe(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, struct mlx5_cqe64 *cqe, int budget_rem) argument
226 mlx5e_decompress_cqes_cont(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, int update_owner_only, int budget_rem) argument
256 mlx5e_decompress_cqes_start(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, int budget_rem) argument
409 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
427 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
441 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
587 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local
764 struct mlx5_wq_cyc *wq = &sq->wq; local
896 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
1082 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local
1824 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
1872 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
1919 struct mlx5_wq_ll *wq; local
2314 struct mlx5_wq_ll *wq; local
2392 struct mlx5_wq_ll *wq; local
2644 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
2734 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
[all...]
/linux-master/drivers/scsi/snic/
H A Dsnic_res.h49 snic_queue_wq_eth_desc(struct vnic_wq *wq, argument
57 struct wq_enet_desc *desc = svnic_wq_next_desc(wq);
72 svnic_wq_post(wq, os_buf, dma_addr, len, 1, 1);
/linux-master/drivers/dma/idxd/
H A Dcompat.c49 struct idxd_wq *wq = confdev_to_wq(dev); local
51 if (is_idxd_wq_kernel(wq))
53 else if (is_idxd_wq_user(wq))
/linux-master/net/sunrpc/
H A Dsched.c320 wait_queue_head_t *wq = bit_waitqueue(m, RPC_TASK_ACTIVE); local
327 spin_lock_irqsave(&wq->lock, flags);
330 if (waitqueue_active(wq))
331 __wake_up_locked_key(wq, TASK_NORMAL, &k);
332 spin_unlock_irqrestore(&wq->lock, flags);
340 * to enforce taking of the wq->lock and hence avoid races with
361 static void rpc_make_runnable(struct workqueue_struct *wq, argument
371 queue_work(wq, &task->u.tk_work);
500 * @wq: workqueue on which to run task
506 static void __rpc_do_wake_up_task_on_wq(struct workqueue_struct *wq, argument
527 rpc_wake_up_task_on_wq_queue_action_locked(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task, bool (*action)(struct rpc_task *, void *), void *data) argument
662 rpc_wake_up_first_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument
1278 struct workqueue_struct *wq; local
1302 struct workqueue_struct *wq = NULL; local
[all...]
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Drx.c23 struct mlx5_wq_cyc *wq = &icosq->wq; local
51 umr_wqe = mlx5_wq_cyc_get_wqe(wq, pi);
162 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
171 contig = mlx5_wq_cyc_get_size(wq) - ix;
181 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i);
186 wqe = mlx5_wq_cyc_get_wqe(wq, j);
200 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
[all...]
/linux-master/fs/netfs/
H A Dlocking.c27 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP);
31 prepare_to_wait(wq, &q.wq_entry, TASK_INTERRUPTIBLE);
38 finish_wait(wq, &q.wq_entry);
/linux-master/drivers/scsi/cxlflash/
H A Docxl_hw.h60 wait_queue_head_t wq; /* Wait queue for poll and interrupts */ member in struct:ocxlflash_context
/linux-master/fs/xfs/
H A Dxfs_pwork.h18 struct workqueue_struct *wq; member in struct:xfs_pwork_ctl
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_events.h64 wait_queue_head_t wq; /* List of event waiters. */ member in struct:kfd_event
/linux-master/drivers/net/ethernet/cavium/liquidio/
H A Dresponse_manager.c42 oct->dma_comp_wq.wq = alloc_workqueue("dma-comp", WQ_MEM_RECLAIM, 0);
43 if (!oct->dma_comp_wq.wq) {
44 dev_err(&oct->pci_dev->dev, "failed to create wq thread\n");
60 destroy_workqueue(oct->dma_comp_wq.wq);
236 queue_delayed_work(cwq->wq, &cwq->wk.work, msecs_to_jiffies(1));
/linux-master/fs/bcachefs/
H A Dclock.h29 __ret = __wait_event_timeout(wq, condition, timeout); \
/linux-master/drivers/soc/apple/
H A Drtkit-internal.h53 struct workqueue_struct *wq; member in struct:apple_rtkit

Completed in 282 milliseconds

1234567891011>>