/linux-master/drivers/net/ethernet/wangxun/txgbe/ |
H A D | txgbe_irq.h | 4 void txgbe_irq_enable(struct wx *wx, bool queues);
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_context.c | 170 * pvr_context_destroy_queues() - Destroy all queues attached to a context. 171 * @ctx: Context to destroy queues on. 174 * It releases all resources attached to the queues bound to this context. 180 pvr_queue_destroy(ctx->queues.fragment); 181 pvr_queue_destroy(ctx->queues.geometry); 184 pvr_queue_destroy(ctx->queues.compute); 187 pvr_queue_destroy(ctx->queues.transfer); 193 * pvr_context_create_queues() - Create all queues attached to a context. 194 * @ctx: Context to create queues on. 210 ctx->queues [all...] |
H A D | pvr_context.h | 64 * @faulty: Set to 1 when the context queues had unfinished job when 72 /** @queues: Union containing all kind of queues. */ 87 } queues; member in struct:pvr_context 95 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL; 97 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.fragment : NULL; 99 return ctx->type == DRM_PVR_CTX_TYPE_COMPUTE ? ctx->queues.compute : NULL; 101 return ctx->type == DRM_PVR_CTX_TYPE_TRANSFER_FRAG ? ctx->queues.transfer : NULL;
|
H A D | pvr_queue.c | 523 job->ctx->queues.fragment); 564 lockdep_assert_held(&pvr_dev->queues.lock); 574 list_move_tail(&queue->node, &pvr_dev->queues.idle); 576 list_move_tail(&queue->node, &pvr_dev->queues.active); 596 mutex_lock(&pvr_dev->queues.lock); 598 mutex_unlock(&pvr_dev->queues.lock); 730 struct pvr_queue *geom_queue = job->ctx->queues.geometry; 731 struct pvr_queue *frag_queue = job->ctx->queues.fragment; 765 /* Make sure we CPU-signal the UFO object, so other queues don't get 819 mutex_lock(&pvr_dev->queues [all...] |
/linux-master/sound/virtio/ |
H A D | virtio_card.h | 47 * @queues: Virtqueue wrappers. 64 struct virtio_snd_queue queues[VIRTIO_SND_VQ_MAX]; member in struct:virtio_snd 86 return &snd->queues[VIRTIO_SND_VQ_CONTROL]; 92 return &snd->queues[VIRTIO_SND_VQ_EVENT]; 98 return &snd->queues[VIRTIO_SND_VQ_TX]; 104 return &snd->queues[VIRTIO_SND_VQ_RX];
|
/linux-master/drivers/net/wireless/silabs/wfx/ |
H A D | queue.c | 233 struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)]; local 239 /* sort the queues */ 243 WARN_ON(num_queues >= ARRAY_SIZE(queues)); 244 queues[num_queues] = &wvif->tx_queue[i]; 246 if (wfx_tx_queue_get_weight(queues[j]) < 247 wfx_tx_queue_get_weight(queues[j - 1])) 248 swap(queues[j - 1], queues[j]); 256 skb = skb_dequeue(&queues[i]->offchan); 264 atomic_inc(&queues[ [all...] |
H A D | data_tx.h | 47 void wfx_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop);
|
/linux-master/drivers/net/wireless/realtek/rtw88/ |
H A D | mac.h | 37 void rtw_mac_flush_queues(struct rtw_dev *rtwdev, u32 queues, bool drop); 42 rtw_mac_flush_queues(rtwdev, BIT(rtwdev->hw->queues) - 1, drop);
|
/linux-master/net/sched/ |
H A D | sch_prio.c | 26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member in struct:prio_sched_data 57 return q->queues[q->prio2band[band & TC_PRIO_MAX]]; 63 return q->queues[q->prio2band[0]]; 65 return q->queues[band]; 103 struct Qdisc *qdisc = q->queues[prio]; 117 struct Qdisc *qdisc = q->queues[prio]; 137 qdisc_reset(q->queues[prio]); 173 qdisc_put(q->queues[prio]); 180 struct Qdisc *queues[TCQ_PRIO_BANDS]; local 198 queues[ [all...] |
H A D | sch_multiq.c | 25 struct Qdisc **queues; member in struct:multiq_sched_data 54 return q->queues[0]; 56 return q->queues[band]; 105 qdisc = q->queues[q->curband]; 137 qdisc = q->queues[curband]; 154 qdisc_reset(q->queues[band]); 166 qdisc_put(q->queues[band]); 168 kfree(q->queues); 196 if (q->queues[i] != &noop_qdisc) { 197 struct Qdisc *child = q->queues[ [all...] |
/linux-master/drivers/nvme/target/ |
H A D | loop.c | 30 struct nvme_loop_queue *queues; member in struct:nvme_loop_ctrl 71 return queue - queue->ctrl->queues; 176 struct nvme_loop_queue *queue = &ctrl->queues[0]; 198 iod->queue = &ctrl->queues[queue_idx]; 222 struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1]; 242 struct nvme_loop_queue *queue = &ctrl->queues[0]; 266 if (!test_and_clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[0].flags)) 268 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); 285 kfree(ctrl->queues); 296 clear_bit(NVME_LOOP_Q_LIVE, &ctrl->queues[ [all...] |
/linux-master/drivers/scsi/aacraid/ |
H A D | comminit.c | 373 struct aac_entry * queues; local 375 struct aac_queue_block * comm = dev->queues; 394 queues = (struct aac_entry *)(((ulong)headers) + hdrsize); 397 comm->queue[HostNormCmdQueue].base = queues; 399 queues += HOST_NORM_CMD_ENTRIES; 403 comm->queue[HostHighCmdQueue].base = queues; 406 queues += HOST_HIGH_CMD_ENTRIES; 410 comm->queue[AdapNormCmdQueue].base = queues; 413 queues += ADAP_NORM_CMD_ENTRIES; 417 comm->queue[AdapHighCmdQueue].base = queues; [all...] |
/linux-master/drivers/media/platform/nxp/imx8-isi/ |
H A D | imx8-isi-m2m.c | 52 /* Protects the m2m vb2 queues */ 58 } queues; member in struct:mxc_isi_m2m_ctx 85 return &ctx->queues.out; 87 return &ctx->queues.cap; 112 src_vbuf->sequence = ctx->queues.out.sequence++; 113 dst_vbuf->sequence = ctx->queues.cap.sequence++; 135 .width = ctx->queues.out.format.width, 136 .height = ctx->queues.out.format.height, 139 .width = ctx->queues.cap.format.width, 140 .height = ctx->queues [all...] |
/linux-master/drivers/target/ |
H A D | target_core_tmr.c | 118 flush_work(&dev->queues[i].sq.work); 120 spin_lock_irqsave(&dev->queues[i].lock, flags); 121 list_for_each_entry_safe(se_cmd, next, &dev->queues[i].state_list, 148 spin_unlock_irqrestore(&dev->queues[i].lock, flags); 163 spin_unlock_irqrestore(&dev->queues[i].lock, flags); 301 flush_work(&dev->queues[i].sq.work); 303 spin_lock_irqsave(&dev->queues[i].lock, flags); 304 list_for_each_entry_safe(cmd, next, &dev->queues[i].state_list, 333 spin_unlock_irqrestore(&dev->queues[i].lock, flags);
|
/linux-master/include/linux/ |
H A D | ptr_ring.h | 625 void ***queues; local 628 queues = kmalloc_array(nrings, sizeof(*queues), gfp); 629 if (!queues) 633 queues[i] = __ptr_ring_init_queue_alloc(size, gfp); 634 if (!queues[i]) 641 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], 648 kvfree(queues[i]); 650 kfree(queues); [all...] |
/linux-master/tools/perf/util/ |
H A D | intel-bts.c | 46 struct auxtrace_queues queues; member in struct:intel_bts 211 for (i = 0; i < bts->queues.nr_queues; i++) { 212 ret = intel_bts_setup_queue(bts, &bts->queues.queue_array[i], 222 if (bts->queues.new_data) { 223 bts->queues.new_data = false; 465 queue = &btsq->bts->queues.queue_array[btsq->queue_nr]; 539 struct auxtrace_queues *queues = &bts->queues; local 542 for (i = 0; i < queues->nr_queues; i++) { 543 struct auxtrace_queue *queue = &bts->queues 710 struct auxtrace_queues *queues = &bts->queues; local [all...] |
H A D | s390-cpumsf.c | 47 * To sort the queues in chronological order, all queue access is controlled 54 * After the auxtrace infrastructure has been setup, the auxtrace queues are 61 * record sample, the auxtrace queues will be processed. As auxtrace queues 170 struct auxtrace_queues queues; member in struct:s390_cpumsf 203 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) 206 q = &sf->queues.queue_array[sample->cpu]; 701 queue = &sfq->sf->queues.queue_array[sfq->queue_nr]; 825 for (i = 0; i < sf->queues.nr_queues; i++) { 826 ret = s390_cpumsf_setup_queue(sf, &sf->queues 1016 struct auxtrace_queues *queues = &sf->queues; local [all...] |
H A D | auxtrace.c | 221 int auxtrace_queues__init(struct auxtrace_queues *queues) argument 223 queues->nr_queues = AUXTRACE_INIT_NR_QUEUES; 224 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); 225 if (!queues->queue_array) 230 static int auxtrace_queues__grow(struct auxtrace_queues *queues, argument 233 unsigned int nr_queues = queues->nr_queues; 243 if (nr_queues < queues->nr_queues || nr_queues < new_nr_queues) 250 for (i = 0; i < queues->nr_queues; i++) { 251 list_splice_tail(&queues 287 auxtrace_queues__queue_buffer(struct auxtrace_queues *queues, unsigned int idx, struct auxtrace_buffer *buffer) argument 321 auxtrace_queues__split_buffer(struct auxtrace_queues *queues, unsigned int idx, struct auxtrace_buffer *buffer) argument 359 auxtrace_queues__add_buffer(struct auxtrace_queues *queues, struct perf_session *session, unsigned int idx, struct auxtrace_buffer *buffer, struct auxtrace_buffer **buffer_ptr) argument 404 auxtrace_queues__add_event(struct auxtrace_queues *queues, struct perf_session *session, union perf_event *event, off_t data_offset, struct auxtrace_buffer **buffer_ptr) argument 424 auxtrace_queues__add_indexed_event(struct auxtrace_queues *queues, struct perf_session *session, off_t file_offset, size_t sz) argument 451 auxtrace_queues__free(struct auxtrace_queues *queues) argument 989 auxtrace_queues__process_index_entry(struct auxtrace_queues *queues, struct perf_session *session, struct auxtrace_index_entry *ent) argument 997 auxtrace_queues__process_index(struct auxtrace_queues *queues, struct perf_session *session) argument 1037 auxtrace_queues__sample_queue(struct auxtrace_queues *queues, struct perf_sample *sample, struct perf_session *session) argument 1061 auxtrace_queues__add_sample(struct auxtrace_queues *queues, struct perf_session *session, struct perf_sample *sample, u64 data_offset, u64 reference) argument [all...] |
H A D | arm-spe.c | 42 struct auxtrace_queues queues; member in struct:arm_spe 154 queue = &speq->spe->queues.queue_array[speq->queue_nr]; 271 arm_spe_set_pid_tid_cpu(spe, &spe->queues.queue_array[speq->queue_nr]); 768 for (i = 0; i < spe->queues.nr_queues; i++) { 769 ret = arm_spe__setup_queue(spe, &spe->queues.queue_array[i], i); 779 if (spe->queues.new_data) { 780 spe->queues.new_data = false; 822 queue = &spe->queues.queue_array[queue_nr]; 863 struct auxtrace_queues *queues = &spe->queues; local 1035 struct auxtrace_queues *queues = &spe->queues; local [all...] |
/linux-master/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_debugfs.c | 126 struct dentry *queues, *tx, *rx, *xdp; local 140 queues = debugfs_create_dir("queue", nn->debugfs_dir); 142 rx = debugfs_create_dir("rx", queues); 143 tx = debugfs_create_dir("tx", queues); 144 xdp = debugfs_create_dir("xdp", queues);
|
/linux-master/drivers/vdpa/virtio_pci/ |
H A D | vp_vdpa.c | 42 int queues; member in struct:vp_vdpa 116 for (i = 0; i < vp_vdpa->queues; i++) { 162 int queues = vp_vdpa->queues; local 163 int vectors = queues + 1; 175 for (i = 0; i < queues; i++) { 194 irq = pci_irq_vector(pdev, queues); 202 vp_modern_config_vector(mdev, queues); 502 vp_vdpa->queues = vp_modern_get_num_queues(mdev); 526 vp_vdpa->vring = devm_kcalloc(&pdev->dev, vp_vdpa->queues, [all...] |
/linux-master/drivers/vdpa/alibaba/ |
H A D | eni_vdpa.c | 45 int queues; member in struct:eni_vdpa 118 for (i = 0; i < eni_vdpa->queues; i++) { 164 int queues = eni_vdpa->queues; local 165 int vectors = queues + 1; 177 for (i = 0; i < queues; i++) { 195 irq = pci_irq_vector(pdev, queues); 202 vp_legacy_config_vector(ldev, queues); 500 eni_vdpa->queues = eni_vdpa_get_num_queues(eni_vdpa); 502 eni_vdpa->vring = devm_kcalloc(&pdev->dev, eni_vdpa->queues, [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | vfio_ap_private.h | 67 * @aqm: identifies the AP queues (domains) in the matrix 72 * queues (domains) and control domains that belong to an AP matrix. The bits in 88 * @queues: a hashtable of queues (struct vfio_ap_queue). 91 DECLARE_HASHTABLE(queues, 8); 106 * @qtable: table of queues (struct vfio_ap_queue) assigned to the mdev 136 * @reset_qnode: allows the vfio_ap_queue struct to be added to a list of queues
|
/linux-master/fs/autofs/ |
H A D | waitq.c | 28 wq = sbi->queues; 29 sbi->queues = NULL; /* Erase all wait queues */ 184 for (wq = sbi->queues; wq; wq = wq->next) { 389 wq->next = sbi->queues; 390 sbi->queues = wq; 493 for (wql = &sbi->queues; (wq = *wql) != NULL; wql = &wq->next) {
|
/linux-master/drivers/net/ethernet/cadence/ |
H A D | macb_main.c | 501 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { 711 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) 765 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) 1097 (unsigned int)(queue - bp->queues), 1235 u16 queue_index = queue - bp->queues; 1682 (unsigned int)(queue - bp->queues), work_done, budget); 1768 (unsigned int)(queue - bp->queues), work_done, budget); 1803 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) { 1821 for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) 1851 (unsigned int)(queue - bp->queues), [all...] |