/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 422 struct lpfc_queue *wq; local 426 wq = phba->sli4_hba.hdwq[wqidx].io_wq; 429 wq = phba->sli4_hba.mbx_wq; 432 wq = phba->sli4_hba.els_wq; 435 wq = phba->sli4_hba.nvmels_wq; 442 qtypestr, wqidx, wq->queue_id); 445 qtypestr, wq->queue_id); 447 lpfc_debug_dump_q(wq); 462 struct lpfc_queue *wq, *cq, *eq; local 466 /* io wq an [all...] |
/linux-master/drivers/char/tpm/ |
H A D | tpm_ibmvtpm.h | 29 wait_queue_head_t wq; member in struct:ibmvtpm_crq_queue 41 wait_queue_head_t wq; member in struct:ibmvtpm_dev
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_qp.h | 54 #define HINIC_MIN_TX_WQE_SIZE(wq) \ 55 ALIGN(HINIC_SQ_WQE_SIZE(1), (wq)->wqebb_size) 58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size) 82 struct hinic_wq *wq; member in struct:hinic_sq 101 struct hinic_wq *wq; member in struct:hinic_rq 139 struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr, 145 struct hinic_wq *wq, struct msix_entry *entry);
|
H A D | hinic_hw_cmdq.c | 343 struct hinic_wq *wq = cmdq->wq; local 350 /* WQE_SIZE = WQEBB_SIZE, we will get the wq element and not shadow*/ 351 hw_wqe = hinic_get_wqe(wq, WQE_LCMD_SIZE, &curr_prod_idx); 361 num_wqebbs = ALIGN(WQE_LCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; 363 if (next_prod_idx >= wq->q_depth) { 365 next_prod_idx -= wq->q_depth; 380 /* CMDQ WQE is not shadow, therefore wqe will be written to wq */ 422 struct hinic_wq *wq local 699 struct hinic_wq *wq = cmdq->wq; local 738 init_cmdq(struct hinic_cmdq *cmdq, struct hinic_wq *wq, enum hinic_cmdq_type q_type, void __iomem *db_area) argument [all...] |
/linux-master/drivers/power/supply/ |
H A D | ipaq_micro_battery.c | 40 struct workqueue_struct *wq; member in struct:micro_battery 88 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); 235 mb->wq = alloc_workqueue("ipaq-battery-wq", WQ_MEM_RECLAIM, 0); 236 if (!mb->wq) 241 queue_delayed_work(mb->wq, &mb->update, 1); 264 destroy_workqueue(mb->wq); 276 destroy_workqueue(mb->wq); 291 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD));
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_guc_submit_types.h | 60 /** @wq: Workqueue info */ 61 u32 wq[WQ_SIZE / sizeof(u32)]; member in struct:guc_submit_parallel_scratch 145 /** @wq: Workqueue Items */ 146 u32 wq[WQ_SIZE / sizeof(u32)]; member in struct:xe_guc_submit_exec_queue_snapshot::__anon626
|
/linux-master/tools/testing/selftests/net/ |
H A D | ioam6.sh | 308 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.ioam6_id=${ALPHA[0]} 309 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.ioam6_id_wide=${ALPHA[1]} 310 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.conf.veth0.ioam6_id=${ALPHA[4]} 311 ip netns exec $ioam_node_alpha sysctl -wq net.ipv6.conf.veth0.ioam6_id_wide=${ALPHA[5]} 316 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.all.forwarding=1 317 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.ioam6_id=${BETA[0]} 318 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.ioam6_id_wide=${BETA[1]} 319 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.veth0.ioam6_enabled=1 320 ip netns exec $ioam_node_beta sysctl -wq net.ipv6.conf.veth0.ioam6_id=${BETA[2]} 321 ip netns exec $ioam_node_beta sysctl -wq ne [all...] |
H A D | srv6_end_dt46_l3vpn_test.sh | 264 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.accept_dad=0 265 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.default.accept_dad=0 271 ip netns exec ${nsname} sysctl -wq net.ipv4.ip_forward=1 272 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.forwarding=1 285 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.all.accept_dad=0 286 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.default.accept_dad=0 290 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.all.rp_filter=0 291 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.default.rp_filter=0 305 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.all.accept_dad=0 306 ip netns exec ${rtname} sysctl -wq ne [all...] |
/linux-master/drivers/infiniband/hw/mana/ |
H A D | qp.c | 110 struct mana_ib_wq *wq; local 192 wq = container_of(ibwq, struct mana_ib_wq, ibwq); 197 wq_spec.gdma_region = wq->gdma_region; 198 wq_spec.queue_size = wq->wq_buf_size; 207 &wq_spec, &cq_spec, &wq->rx_object); 215 wq->gdma_region = GDMA_INVALID_DMA_REGION; 218 wq->id = wq_spec.queue_index; 222 "ret %d rx_object 0x%llx wq id %llu cq id %llu\n", 223 ret, wq->rx_object, wq 477 struct mana_ib_wq *wq; local [all...] |
/linux-master/drivers/hid/ |
H A D | hid-elo.c | 34 static struct workqueue_struct *wq; variable in typeref:struct:workqueue_struct 182 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); 258 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); 296 wq = create_singlethread_workqueue("elousb"); 297 if (!wq) 302 destroy_workqueue(wq); 311 destroy_workqueue(wq);
|
/linux-master/drivers/i2c/busses/ |
H A D | i2c-taos-evm.c | 30 static DECLARE_WAIT_QUEUE_HEAD(wq); 104 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE, 161 wake_up_interruptible(&wq); 166 wake_up_interruptible(&wq); 173 wake_up_interruptible(&wq); 226 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE, 248 wait_event_interruptible_timeout(wq, taos->state == TAOS_STATE_IDLE,
|
/linux-master/block/ |
H A D | blk-rq-qos.c | 198 struct wait_queue_entry wq; member in struct:rq_qos_wait_data 211 wq); 248 .wq = { 250 .entry = LIST_HEAD_INIT(data.wq.entry), 263 has_sleeper = !prepare_to_wait_exclusive(&rqw->wait, &data.wq, 270 finish_wait(&rqw->wait, &data.wq); 286 finish_wait(&rqw->wait, &data.wq);
|
/linux-master/drivers/platform/x86/ |
H A D | asus-wireless.c | 27 struct workqueue_struct *wq; member in struct:asus_wireless_data 108 queue_work(data->wq, &data->led_work); 157 data->wq = create_singlethread_workqueue("asus_wireless_workqueue"); 158 if (!data->wq) 169 destroy_workqueue(data->wq); 178 if (data->wq) { 180 destroy_workqueue(data->wq);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_rx.c | 81 static inline void mlx5e_read_cqe_slot(struct mlx5_cqwq *wq, argument 84 u32 ci = mlx5_cqwq_ctr2ix(wq, cqcc); 86 memcpy(data, mlx5_cqwq_get_wqe(wq, ci), sizeof(struct mlx5_cqe64)); 105 mlx5_wq_cyc_ctr2ix(&rq->wqe.wq, be16_to_cpu(title->wqe_counter) + 1); 109 struct mlx5_cqwq *wq, 115 mlx5e_read_cqe_slot(wq, cqcc, title); 121 static inline void mlx5e_read_mini_arr_slot(struct mlx5_cqwq *wq, argument 125 mlx5e_read_cqe_slot(wq, cqcc, cqd->mini_arr); 129 static inline void mlx5e_cqes_update_owner(struct mlx5_cqwq *wq, int n) argument 131 u32 cqcc = wq 108 mlx5e_read_title_slot(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument 153 mlx5e_decompress_cqe(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument 183 mlx5e_decompress_cqe_no_hash(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, u32 cqcc) argument 194 mlx5e_decompress_enhanced_cqe(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, struct mlx5_cqe64 *cqe, int budget_rem) argument 226 mlx5e_decompress_cqes_cont(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, int update_owner_only, int budget_rem) argument 256 mlx5e_decompress_cqes_start(struct mlx5e_rq *rq, struct mlx5_cqwq *wq, int budget_rem) argument 409 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 427 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 441 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 587 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local 764 struct mlx5_wq_cyc *wq = &sq->wq; local 896 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 1082 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local 1824 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 1872 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 1919 struct mlx5_wq_ll *wq; local 2314 struct mlx5_wq_ll *wq; local 2392 struct mlx5_wq_ll *wq; local 2644 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 2734 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local [all...] |
/linux-master/drivers/scsi/snic/ |
H A D | snic_res.h | 49 snic_queue_wq_eth_desc(struct vnic_wq *wq, argument 57 struct wq_enet_desc *desc = svnic_wq_next_desc(wq); 72 svnic_wq_post(wq, os_buf, dma_addr, len, 1, 1);
|
/linux-master/drivers/dma/idxd/ |
H A D | compat.c | 49 struct idxd_wq *wq = confdev_to_wq(dev); local 51 if (is_idxd_wq_kernel(wq)) 53 else if (is_idxd_wq_user(wq))
|
/linux-master/net/sunrpc/ |
H A D | sched.c | 320 wait_queue_head_t *wq = bit_waitqueue(m, RPC_TASK_ACTIVE); local 327 spin_lock_irqsave(&wq->lock, flags); 330 if (waitqueue_active(wq)) 331 __wake_up_locked_key(wq, TASK_NORMAL, &k); 332 spin_unlock_irqrestore(&wq->lock, flags); 340 * to enforce taking of the wq->lock and hence avoid races with 361 static void rpc_make_runnable(struct workqueue_struct *wq, argument 371 queue_work(wq, &task->u.tk_work); 500 * @wq: workqueue on which to run task 506 static void __rpc_do_wake_up_task_on_wq(struct workqueue_struct *wq, argument 527 rpc_wake_up_task_on_wq_queue_action_locked(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task, bool (*action)(struct rpc_task *, void *), void *data) argument 662 rpc_wake_up_first_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument 1278 struct workqueue_struct *wq; local 1302 struct workqueue_struct *wq = NULL; local [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | rx.c | 23 struct mlx5_wq_cyc *wq = &icosq->wq; local 51 umr_wqe = mlx5_wq_cyc_get_wqe(wq, pi); 162 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 171 contig = mlx5_wq_cyc_get_size(wq) - ix; 181 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); 186 wqe = mlx5_wq_cyc_get_wqe(wq, j); 200 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local [all...] |
/linux-master/fs/netfs/ |
H A D | locking.c | 27 wait_queue_head_t *wq = bit_waitqueue(&inode->i_state, __I_DIO_WAKEUP); 31 prepare_to_wait(wq, &q.wq_entry, TASK_INTERRUPTIBLE); 38 finish_wait(wq, &q.wq_entry);
|
/linux-master/drivers/scsi/cxlflash/ |
H A D | ocxl_hw.h | 60 wait_queue_head_t wq; /* Wait queue for poll and interrupts */ member in struct:ocxlflash_context
|
/linux-master/fs/xfs/ |
H A D | xfs_pwork.h | 18 struct workqueue_struct *wq; member in struct:xfs_pwork_ctl
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_events.h | 64 wait_queue_head_t wq; /* List of event waiters. */ member in struct:kfd_event
|
/linux-master/drivers/net/ethernet/cavium/liquidio/ |
H A D | response_manager.c | 42 oct->dma_comp_wq.wq = alloc_workqueue("dma-comp", WQ_MEM_RECLAIM, 0); 43 if (!oct->dma_comp_wq.wq) { 44 dev_err(&oct->pci_dev->dev, "failed to create wq thread\n"); 60 destroy_workqueue(oct->dma_comp_wq.wq); 236 queue_delayed_work(cwq->wq, &cwq->wk.work, msecs_to_jiffies(1));
|
/linux-master/fs/bcachefs/ |
H A D | clock.h | 29 __ret = __wait_event_timeout(wq, condition, timeout); \
|
/linux-master/drivers/soc/apple/ |
H A D | rtkit-internal.h | 53 struct workqueue_struct *wq; member in struct:apple_rtkit
|