Searched refs:wq (Results 126 - 150 of 700) sorted by relevance

1234567891011>>

/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c149 queue_work(priv->wq, &priv->update_stats_work);
163 queue_work(priv->wq, &priv->update_carrier_work);
357 int wq_sz = mlx5_wq_ll_get_size(&rq->mpwqe.wq);
541 u32 wq_size = mlx5_wq_ll_get_size(&rq->mpwqe.wq);
590 for (i = 0; i < mlx5_wq_cyc_get_size(&rq->wqe.wq); i++) {
628 for (i = 0; i < mlx5_wq_cyc_get_size(&rq->wqe.wq); i++) {
640 int wq_sz = mlx5_wq_cyc_get_size(&rq->wqe.wq);
743 void *wqc = MLX5_ADDR_OF(rqc, rqp->rqc, wq);
769 wq_size = BIT(MLX5_GET(wq, wqc, log_wq_sz));
802 void *rqc_wq = MLX5_ADDR_OF(rqc, rqc, wq);
1034 void *wq; local
1199 struct mlx5_wq_ll *wq; local
1239 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local
1258 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local
1426 struct mlx5_wq_cyc *wq = &sq->wq; local
1513 struct mlx5_wq_cyc *wq = &sq->wq; local
1590 struct mlx5_wq_cyc *wq = &sq->wq; local
1651 void *wq; local
1823 struct mlx5_wq_cyc *wq = &sq->wq; local
[all...]
H A Den_tx.c344 if (unlikely(!mlx5e_wqc_has_room_for(&sq->wq, sq->cc, sq->pc, sq->stop_room))) {
359 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc);
366 wqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc);
367 mlx5e_notify_hw(&sq->wq, sq->pc, sq->uar_map, &wqe->ctrl);
377 struct mlx5_wq_cyc *wq = &sq->wq; local
419 mlx5e_notify_hw(wq, sq->pc, sq->uar_map, cseg);
578 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc);
630 mlx5e_notify_hw(&sq->wq, sq->pc, sq->uar_map, cseg);
635 mlx5e_notify_hw(&sq->wq, s
[all...]
H A Dfw_reset.c21 struct workqueue_struct *wq; member in struct:mlx5_fw_reset
274 queue_work(fw_reset->wq, &fw_reset->reset_reload_work);
640 queue_work(fw_reset->wq, &fw_reset->reset_request_work);
643 queue_work(fw_reset->wq, &fw_reset->reset_unload_work);
646 queue_work(fw_reset->wq, &fw_reset->reset_now_work);
649 queue_work(fw_reset->wq, &fw_reset->reset_abort_work);
664 queue_work(fw_reset->wq, &fw_reset->fw_live_patch_work);
756 fw_reset->wq = create_singlethread_workqueue("mlx5_fw_reset_events");
757 if (!fw_reset->wq) {
769 destroy_workqueue(fw_reset->wq);
[all...]
/linux-master/drivers/media/pci/netup_unidvb/
H A Dnetup_unidvb.h82 wait_queue_head_t wq; member in struct:netup_i2c
118 struct workqueue_struct *wq; member in struct:netup_unidvb_dev
/linux-master/drivers/md/
H A Ddm-zoned-reclaim.c18 struct workqueue_struct *wq; member in struct:dmz_reclaim
514 mod_delayed_work(zrc->wq, &zrc->work, DMZ_IDLE_PERIOD);
577 zrc->wq = alloc_ordered_workqueue("dmz_rwq_%s_%d", WQ_MEM_RECLAIM,
579 if (!zrc->wq) {
585 queue_delayed_work(zrc->wq, &zrc->work, 0);
602 destroy_workqueue(zrc->wq);
620 queue_delayed_work(zrc->wq, &zrc->work, DMZ_IDLE_PERIOD);
639 mod_delayed_work(zrc->wq, &zrc->work, 0);
H A Ddm-bio-prison-v2.h57 struct dm_bio_prison_v2 *dm_bio_prison_create_v2(struct workqueue_struct *wq);
/linux-master/drivers/infiniband/sw/rdmavt/
H A Drc.c72 head = RDMA_READ_UAPI_ATOMIC(qp->r_rq.wq->head);
73 tail = RDMA_READ_UAPI_ATOMIC(qp->r_rq.wq->tail);
/linux-master/drivers/gpu/drm/
H A Ddrm_flip_work.c81 * @wq: the work-queue to run the queued work on
89 struct workqueue_struct *wq)
97 queue_work(wq, &work->worker);
88 drm_flip_work_commit(struct drm_flip_work *work, struct workqueue_struct *wq) argument
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dtx.c49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc);
56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc);
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_reset.h84 struct workqueue_struct *wq; member in struct:amdgpu_reset_domain
137 return queue_work(domain->wq, work);
/linux-master/drivers/net/ethernet/cisco/enic/
H A Denic_main.c324 static void enic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument
326 struct enic *enic = vnic_dev_priv(wq->vdev);
339 static void enic_wq_free_buf(struct vnic_wq *wq, argument
342 enic_free_wq_buf(wq, buf);
352 vnic_wq_service(&enic->wq[q_number], cq_desc,
357 vnic_wq_desc_avail(&enic->wq[q_number]) >=
373 error_status = vnic_wq_error_status(&enic->wq[i]);
544 static int enic_queue_wq_skb_cont(struct enic *enic, struct vnic_wq *wq, argument
559 enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag),
567 static int enic_queue_wq_skb_vlan(struct enic *enic, struct vnic_wq *wq, argument
596 enic_queue_wq_skb_csum_l4(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
668 enic_queue_wq_skb_tso(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int mss, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
738 enic_queue_wq_skb_encap(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument
770 enic_queue_wq_skb(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb) argument
824 struct vnic_wq *wq; local
1607 struct vnic_wq *wq = &enic->wq[wq_index]; local
1756 int wq = enic_cq_wq(enic, i); local
[all...]
/linux-master/drivers/scsi/snic/
H A Dvnic_dev.c23 struct vnic_wq wq; member in struct:devcmd2_controller
275 pr_err("%s: wq is full while issuing devcmd2 command %d, fetch index: %u, posted index: %u\n",
365 &dc2c->wq,
371 fetch_idx = ioread32(&dc2c->wq.ctrl->fetch_index);
382 vnic_wq_init_start(&dc2c->wq, 0, fetch_idx, fetch_idx, 0, 0);
383 svnic_wq_enable(&dc2c->wq);
392 dc2c->cmd_ring = (struct vnic_devcmd2 *) dc2c->wq.ring.descs;
393 dc2c->wq_ctrl = dc2c->wq.ctrl;
410 svnic_wq_disable(&dc2c->wq);
411 svnic_wq_free(&dc2c->wq);
[all...]
H A Dsnic_res.c85 SNIC_INFO("vNIC resources wq %d\n", c->wq_enet_desc_count);
123 svnic_wq_free(&snic->wq[i]);
156 SNIC_INFO("wq %d cq %d intr %d\n", snic->wq_count,
164 &snic->wq[i],
210 svnic_wq_init(&snic->wq[i],
274 err_status = ioread32(&snic->wq[i].ctrl->error_status);
/linux-master/tools/testing/selftests/net/
H A Dsrv6_end_next_csid_l3vpn_test.sh518 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0
519 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0
520 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.forwarding=1
522 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.all.rp_filter=0
523 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.default.rp_filter=0
524 ip netns exec "${nsname}" sysctl -wq net.ipv4.ip_forward=1
750 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0
751 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0
790 sysctl -wq net.ipv6.conf."${RT2HS_DEVNAME}".proxy_ndp=1
792 sysctl -wq ne
[all...]
H A Dsrv6_end_x_next_csid_l3vpn_test.sh458 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0
459 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0
460 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.forwarding=1
462 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.all.rp_filter=0
463 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.default.rp_filter=0
464 ip netns exec "${nsname}" sysctl -wq net.ipv4.ip_forward=1
810 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0
811 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0
850 sysctl -wq net.ipv6.conf."${RT2HS_DEVNAME}".proxy_ndp=1
852 sysctl -wq ne
[all...]
/linux-master/fs/
H A Duserfaultfd.c68 wait_queue_entry_t wq; member in struct:userfaultfd_wait_queue
121 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, argument
129 uwq = container_of(wq, struct userfaultfd_wait_queue, wq);
142 ret = wake_up_state(wq->private, mode);
152 * wq->private is read before calling the extern function
155 list_del_init(&wq->entry);
484 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function);
485 uwq.wq.private = current;
507 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq);
934 wait_queue_entry_t *wq; local
2145 wait_queue_entry_t *wq; local
[all...]
/linux-master/drivers/scsi/elx/efct/
H A Defct_hw.c338 io->wq = NULL;
1087 efct_hw_queue_hash_add(hw->wq_hash, hw->wq[i].id, i);
1591 io->wq = hw->wq_cpu_array[raw_smp_processor_id()];
1592 if (!io->wq) {
1595 io->wq = hw->hw_wq[0];
1958 if (!io_to_abort->wq) {
2012 if (io_to_abort->wq) {
2013 spin_lock_irqsave(&io_to_abort->wq->queue->lock, flags);
2016 spin_unlock_irqrestore(&io_to_abort->wq->queue->lock,
2020 spin_unlock_irqrestore(&io_to_abort->wq
2260 _efct_hw_wq_write(struct hw_wq *wq, struct efct_hw_wqe *wqe) argument
2284 hw_wq_submit_pending(struct hw_wq *wq, u32 update_free_count) argument
2365 struct hw_wq *wq = NULL; local
2511 efct_hw_wq_write(struct hw_wq *wq, struct efct_hw_wqe *wqe) argument
2978 struct hw_wq *wq; local
[all...]
/linux-master/include/linux/power/
H A Dcharger-manager.h45 * @wq: the workqueue to control charger according to the state of
63 struct work_struct wq; member in struct:charger_cable
/linux-master/drivers/gpu/drm/omapdrm/
H A Domap_irq.c13 wait_queue_head_t wq; member in struct:omap_irq_wait
38 wake_up(&wait->wq);
48 init_waitqueue_head(&wait->wq);
67 ret = wait_event_timeout(wait->wq, (wait->count <= 0), timeout);
H A Domap_drv.h79 struct workqueue_struct *wq; member in struct:omap_drm_private
/linux-master/include/drm/
H A Ddrm_suballoc.h17 * @wq: Wait queue for sleeping allocations on contention.
25 wait_queue_head_t wq; member in struct:drm_suballoc_manager
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_eswitch_br.h75 struct workqueue_struct *wq; member in struct:ice_esw_br_offloads
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_gsc_uc.h44 struct workqueue_struct *wq; member in struct:intel_gsc_uc
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dhealth.c33 devlink_fmsg_u32_pair_put(fmsg, "ci", mlx5_cqwq_get_ci(&cq->wq));
34 devlink_fmsg_u32_pair_put(fmsg, "size", mlx5_cqwq_get_size(&cq->wq));
43 cq_sz = mlx5_cqwq_get_size(&cq->wq);
44 cq_log_stride = mlx5_cqwq_get_log_stride_size(&cq->wq);
/linux-master/drivers/infiniband/hw/hfi1/
H A Diowait.h145 * @wq: workqueue for schedule
149 struct workqueue_struct *wq, int cpu)
151 return !!queue_work_on(cpu, wq, &wait->wait[IOWAIT_IB_SE].iowork);
157 * @wq: the work queue
161 struct workqueue_struct *wq, int cpu)
163 return !!queue_work_on(cpu, wq, &wait->wait[IOWAIT_TID_SE].iowork);
148 iowait_schedule(struct iowait *wait, struct workqueue_struct *wq, int cpu) argument
160 iowait_tid_schedule(struct iowait *wait, struct workqueue_struct *wq, int cpu) argument

Completed in 225 milliseconds

1234567891011>>