/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 149 queue_work(priv->wq, &priv->update_stats_work); 163 queue_work(priv->wq, &priv->update_carrier_work); 357 int wq_sz = mlx5_wq_ll_get_size(&rq->mpwqe.wq); 541 u32 wq_size = mlx5_wq_ll_get_size(&rq->mpwqe.wq); 590 for (i = 0; i < mlx5_wq_cyc_get_size(&rq->wqe.wq); i++) { 628 for (i = 0; i < mlx5_wq_cyc_get_size(&rq->wqe.wq); i++) { 640 int wq_sz = mlx5_wq_cyc_get_size(&rq->wqe.wq); 743 void *wqc = MLX5_ADDR_OF(rqc, rqp->rqc, wq); 769 wq_size = BIT(MLX5_GET(wq, wqc, log_wq_sz)); 802 void *rqc_wq = MLX5_ADDR_OF(rqc, rqc, wq); 1034 void *wq; local 1199 struct mlx5_wq_ll *wq; local 1239 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; local 1258 struct mlx5_wq_cyc *wq = &rq->wqe.wq; local 1426 struct mlx5_wq_cyc *wq = &sq->wq; local 1513 struct mlx5_wq_cyc *wq = &sq->wq; local 1590 struct mlx5_wq_cyc *wq = &sq->wq; local 1651 void *wq; local 1823 struct mlx5_wq_cyc *wq = &sq->wq; local [all...] |
H A D | en_tx.c | 344 if (unlikely(!mlx5e_wqc_has_room_for(&sq->wq, sq->cc, sq->pc, sq->stop_room))) { 359 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); 366 wqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); 367 mlx5e_notify_hw(&sq->wq, sq->pc, sq->uar_map, &wqe->ctrl); 377 struct mlx5_wq_cyc *wq = &sq->wq; local 419 mlx5e_notify_hw(wq, sq->pc, sq->uar_map, cseg); 578 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); 630 mlx5e_notify_hw(&sq->wq, sq->pc, sq->uar_map, cseg); 635 mlx5e_notify_hw(&sq->wq, s [all...] |
H A D | fw_reset.c | 21 struct workqueue_struct *wq; member in struct:mlx5_fw_reset 274 queue_work(fw_reset->wq, &fw_reset->reset_reload_work); 640 queue_work(fw_reset->wq, &fw_reset->reset_request_work); 643 queue_work(fw_reset->wq, &fw_reset->reset_unload_work); 646 queue_work(fw_reset->wq, &fw_reset->reset_now_work); 649 queue_work(fw_reset->wq, &fw_reset->reset_abort_work); 664 queue_work(fw_reset->wq, &fw_reset->fw_live_patch_work); 756 fw_reset->wq = create_singlethread_workqueue("mlx5_fw_reset_events"); 757 if (!fw_reset->wq) { 769 destroy_workqueue(fw_reset->wq); [all...] |
/linux-master/drivers/media/pci/netup_unidvb/ |
H A D | netup_unidvb.h | 82 wait_queue_head_t wq; member in struct:netup_i2c 118 struct workqueue_struct *wq; member in struct:netup_unidvb_dev
|
/linux-master/drivers/md/ |
H A D | dm-zoned-reclaim.c | 18 struct workqueue_struct *wq; member in struct:dmz_reclaim 514 mod_delayed_work(zrc->wq, &zrc->work, DMZ_IDLE_PERIOD); 577 zrc->wq = alloc_ordered_workqueue("dmz_rwq_%s_%d", WQ_MEM_RECLAIM, 579 if (!zrc->wq) { 585 queue_delayed_work(zrc->wq, &zrc->work, 0); 602 destroy_workqueue(zrc->wq); 620 queue_delayed_work(zrc->wq, &zrc->work, DMZ_IDLE_PERIOD); 639 mod_delayed_work(zrc->wq, &zrc->work, 0);
|
H A D | dm-bio-prison-v2.h | 57 struct dm_bio_prison_v2 *dm_bio_prison_create_v2(struct workqueue_struct *wq);
|
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | rc.c | 72 head = RDMA_READ_UAPI_ATOMIC(qp->r_rq.wq->head); 73 tail = RDMA_READ_UAPI_ATOMIC(qp->r_rq.wq->tail);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_flip_work.c | 81 * @wq: the work-queue to run the queued work on 89 struct workqueue_struct *wq) 97 queue_work(wq, &work->worker); 88 drm_flip_work_commit(struct drm_flip_work *work, struct workqueue_struct *wq) argument
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | tx.c | 49 u16 pi = mlx5_wq_cyc_ctr2ix(&sq->wq, sq->pc); 56 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc);
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_reset.h | 84 struct workqueue_struct *wq; member in struct:amdgpu_reset_domain 137 return queue_work(domain->wq, work);
|
/linux-master/drivers/net/ethernet/cisco/enic/ |
H A D | enic_main.c | 324 static void enic_free_wq_buf(struct vnic_wq *wq, struct vnic_wq_buf *buf) argument 326 struct enic *enic = vnic_dev_priv(wq->vdev); 339 static void enic_wq_free_buf(struct vnic_wq *wq, argument 342 enic_free_wq_buf(wq, buf); 352 vnic_wq_service(&enic->wq[q_number], cq_desc, 357 vnic_wq_desc_avail(&enic->wq[q_number]) >= 373 error_status = vnic_wq_error_status(&enic->wq[i]); 544 static int enic_queue_wq_skb_cont(struct enic *enic, struct vnic_wq *wq, argument 559 enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag), 567 static int enic_queue_wq_skb_vlan(struct enic *enic, struct vnic_wq *wq, argument 596 enic_queue_wq_skb_csum_l4(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 668 enic_queue_wq_skb_tso(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, unsigned int mss, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 738 enic_queue_wq_skb_encap(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb, int vlan_tag_insert, unsigned int vlan_tag, int loopback) argument 770 enic_queue_wq_skb(struct enic *enic, struct vnic_wq *wq, struct sk_buff *skb) argument 824 struct vnic_wq *wq; local 1607 struct vnic_wq *wq = &enic->wq[wq_index]; local 1756 int wq = enic_cq_wq(enic, i); local [all...] |
/linux-master/drivers/scsi/snic/ |
H A D | vnic_dev.c | 23 struct vnic_wq wq; member in struct:devcmd2_controller 275 pr_err("%s: wq is full while issuing devcmd2 command %d, fetch index: %u, posted index: %u\n", 365 &dc2c->wq, 371 fetch_idx = ioread32(&dc2c->wq.ctrl->fetch_index); 382 vnic_wq_init_start(&dc2c->wq, 0, fetch_idx, fetch_idx, 0, 0); 383 svnic_wq_enable(&dc2c->wq); 392 dc2c->cmd_ring = (struct vnic_devcmd2 *) dc2c->wq.ring.descs; 393 dc2c->wq_ctrl = dc2c->wq.ctrl; 410 svnic_wq_disable(&dc2c->wq); 411 svnic_wq_free(&dc2c->wq); [all...] |
H A D | snic_res.c | 85 SNIC_INFO("vNIC resources wq %d\n", c->wq_enet_desc_count); 123 svnic_wq_free(&snic->wq[i]); 156 SNIC_INFO("wq %d cq %d intr %d\n", snic->wq_count, 164 &snic->wq[i], 210 svnic_wq_init(&snic->wq[i], 274 err_status = ioread32(&snic->wq[i].ctrl->error_status);
|
/linux-master/tools/testing/selftests/net/ |
H A D | srv6_end_next_csid_l3vpn_test.sh | 518 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0 519 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0 520 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.forwarding=1 522 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.all.rp_filter=0 523 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.default.rp_filter=0 524 ip netns exec "${nsname}" sysctl -wq net.ipv4.ip_forward=1 750 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0 751 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0 790 sysctl -wq net.ipv6.conf."${RT2HS_DEVNAME}".proxy_ndp=1 792 sysctl -wq ne [all...] |
H A D | srv6_end_x_next_csid_l3vpn_test.sh | 458 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0 459 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0 460 ip netns exec "${nsname}" sysctl -wq net.ipv6.conf.all.forwarding=1 462 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.all.rp_filter=0 463 ip netns exec "${nsname}" sysctl -wq net.ipv4.conf.default.rp_filter=0 464 ip netns exec "${nsname}" sysctl -wq net.ipv4.ip_forward=1 810 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.all.accept_dad=0 811 ip netns exec "${hsname}" sysctl -wq net.ipv6.conf.default.accept_dad=0 850 sysctl -wq net.ipv6.conf."${RT2HS_DEVNAME}".proxy_ndp=1 852 sysctl -wq ne [all...] |
/linux-master/fs/ |
H A D | userfaultfd.c | 68 wait_queue_entry_t wq; member in struct:userfaultfd_wait_queue 121 static int userfaultfd_wake_function(wait_queue_entry_t *wq, unsigned mode, argument 129 uwq = container_of(wq, struct userfaultfd_wait_queue, wq); 142 ret = wake_up_state(wq->private, mode); 152 * wq->private is read before calling the extern function 155 list_del_init(&wq->entry); 484 init_waitqueue_func_entry(&uwq.wq, userfaultfd_wake_function); 485 uwq.wq.private = current; 507 __add_wait_queue(&ctx->fault_pending_wqh, &uwq.wq); 934 wait_queue_entry_t *wq; local 2145 wait_queue_entry_t *wq; local [all...] |
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_hw.c | 338 io->wq = NULL; 1087 efct_hw_queue_hash_add(hw->wq_hash, hw->wq[i].id, i); 1591 io->wq = hw->wq_cpu_array[raw_smp_processor_id()]; 1592 if (!io->wq) { 1595 io->wq = hw->hw_wq[0]; 1958 if (!io_to_abort->wq) { 2012 if (io_to_abort->wq) { 2013 spin_lock_irqsave(&io_to_abort->wq->queue->lock, flags); 2016 spin_unlock_irqrestore(&io_to_abort->wq->queue->lock, 2020 spin_unlock_irqrestore(&io_to_abort->wq 2260 _efct_hw_wq_write(struct hw_wq *wq, struct efct_hw_wqe *wqe) argument 2284 hw_wq_submit_pending(struct hw_wq *wq, u32 update_free_count) argument 2365 struct hw_wq *wq = NULL; local 2511 efct_hw_wq_write(struct hw_wq *wq, struct efct_hw_wqe *wqe) argument 2978 struct hw_wq *wq; local [all...] |
/linux-master/include/linux/power/ |
H A D | charger-manager.h | 45 * @wq: the workqueue to control charger according to the state of 63 struct work_struct wq; member in struct:charger_cable
|
/linux-master/drivers/gpu/drm/omapdrm/ |
H A D | omap_irq.c | 13 wait_queue_head_t wq; member in struct:omap_irq_wait 38 wake_up(&wait->wq); 48 init_waitqueue_head(&wait->wq); 67 ret = wait_event_timeout(wait->wq, (wait->count <= 0), timeout);
|
H A D | omap_drv.h | 79 struct workqueue_struct *wq; member in struct:omap_drm_private
|
/linux-master/include/drm/ |
H A D | drm_suballoc.h | 17 * @wq: Wait queue for sleeping allocations on contention. 25 wait_queue_head_t wq; member in struct:drm_suballoc_manager
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_eswitch_br.h | 75 struct workqueue_struct *wq; member in struct:ice_esw_br_offloads
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_gsc_uc.h | 44 struct workqueue_struct *wq; member in struct:intel_gsc_uc
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | health.c | 33 devlink_fmsg_u32_pair_put(fmsg, "ci", mlx5_cqwq_get_ci(&cq->wq)); 34 devlink_fmsg_u32_pair_put(fmsg, "size", mlx5_cqwq_get_size(&cq->wq)); 43 cq_sz = mlx5_cqwq_get_size(&cq->wq); 44 cq_log_stride = mlx5_cqwq_get_log_stride_size(&cq->wq);
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | iowait.h | 145 * @wq: workqueue for schedule 149 struct workqueue_struct *wq, int cpu) 151 return !!queue_work_on(cpu, wq, &wait->wait[IOWAIT_IB_SE].iowork); 157 * @wq: the work queue 161 struct workqueue_struct *wq, int cpu) 163 return !!queue_work_on(cpu, wq, &wait->wait[IOWAIT_TID_SE].iowork); 148 iowait_schedule(struct iowait *wait, struct workqueue_struct *wq, int cpu) argument 160 iowait_tid_schedule(struct iowait *wait, struct workqueue_struct *wq, int cpu) argument
|