Searched refs:queues (Results 51 - 75 of 217) sorted by relevance

123456789

/linux-master/drivers/crypto/hisilicon/sec/
H A Dsec_drv.c639 if (!info->queues[i].in_use) {
640 info->queues[i].in_use = true;
644 return &info->queues[i];
976 /* Same QoS for all queues */
1236 ret = sec_queue_config(info, &info->queues[i], i);
1240 ret = sec_queue_irq_init(&info->queues[i]);
1242 sec_queue_unconfig(info, &info->queues[i]);
1266 sec_queue_irq_uninit(&info->queues[j]);
1267 sec_queue_unconfig(info, &info->queues[j]);
1285 sec_queue_irq_uninit(&info->queues[
[all...]
/linux-master/drivers/scsi/arm/
H A Dacornscsi.c700 SCpnt = queue_remove_exclude(&host->queues.issue, host->busyluns);
708 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt);
1746 if (!ok && queue_probetgtlun(&host->queues.disconnected, target, lun))
1761 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt);
1788 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt);
1798 host->SCpnt = queue_remove_tgtluntag(&host->queues.disconnected,
2407 * Purpose : queues a SCSI command
2435 host->stats.queues += 1;
2440 if (!queue_add_cmd_ordered(&host->queues.issue, SCpnt)) {
2467 if (queue_remove_cmd(&host->queues
[all...]
/linux-master/drivers/nvme/host/
H A Dtcp.c183 struct nvme_tcp_queue *queues; member in struct:nvme_tcp_ctrl
213 return queue - queue->ctrl->queues;
520 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx];
541 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1];
551 struct nvme_tcp_queue *queue = &ctrl->queues[0];
1345 struct nvme_tcp_queue *queue = &ctrl->queues[0];
1355 async->queue = &ctrl->queues[0];
1362 struct nvme_tcp_queue *queue = &ctrl->queues[qid];
1650 struct nvme_tcp_queue *queue = &ctrl->queues[qid];
1824 struct nvme_tcp_queue *queue = &ctrl->queues[qi
[all...]
H A Dfc.c155 struct nvme_fc_queue *queues; member in struct:nvme_fc_ctrl
1404 * queues) that are part of the association. E.g. things are torn
1944 * allocates a single tag set for all io queues and sizes
1945 * the io queues to fully hold all possible tags. Thus, the
2147 struct nvme_fc_queue *queue = &ctrl->queues[queue_idx];
2180 ret = __nvme_fc_init_request(ctrl, &ctrl->queues[0],
2219 struct nvme_fc_queue *queue = &ctrl->queues[qidx];
2244 queue = &ctrl->queues[idx];
2309 nvme_fc_free_queue(&ctrl->queues[i]);
2329 struct nvme_fc_queue *queue = &ctrl->queues[ctr
[all...]
H A Dpci.c97 "Number of queues to use for writes. If not set, reads and writes "
102 MODULE_PARM_DESC(poll_queues, "Number of queues to use for polled IO.");
119 struct nvme_queue *queues; member in struct:nvme_dev
193 /* only used for poll queues: */
342 nvme_dbbuf_free(&dev->queues[i]);
399 struct nvme_queue *nvmeq = &dev->queues[0];
412 struct nvme_queue *nvmeq = &dev->queues[hctx_idx + 1];
1133 struct nvme_queue *nvmeq = &dev->queues[0];
1187 * set. Since URGENT priority is zeroes, it makes all queues
1430 nvme_free_queue(&dev->queues[
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_debug.c51 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
160 list_for_each_entry(pqn, &pqm->queues,
222 list_for_each_entry(pqn, &pqm->queues,
327 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
336 list_for_each_entry(pqn, &pqm->queues, process_queue_list)
565 pr_debug("Resumed %d queues\n", resume_count);
624 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
1103 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
H A Dkfd_process_queue_manager.c37 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
71 pr_info("Cannot open more queues for process with pasid 0x%x\n",
166 INIT_LIST_HEAD(&pqm->queues);
209 list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) {
290 * On GFX 9.4.3, increase the number of queues that
306 * for debug process, verify that it is within the static queues limit
339 /* SDMA queues are always allocated statically no matter
420 list_add(&pqn->process_queue_list, &pqm->queues);
436 /* check if queues list is empty unregister process from device */
643 list_for_each_entry(pqn, &pqm->queues, process_queue_lis
[all...]
H A Dkfd_packet_manager.c122 struct list_head *queues,
145 pr_debug("Building runlist ib process count: %d queues count %d\n",
149 list_for_each_entry(cur, queues, list) {
206 pr_debug("Finished map process and queues to runlist\n");
121 pm_create_runlist_ib(struct packet_manager *pm, struct list_head *queues, uint64_t *rl_gpu_addr, size_t *rl_size_bytes) argument
/linux-master/drivers/net/wireless/realtek/rtw88/
H A Dmac.c992 static u32 get_priority_queues(struct rtw_dev *rtwdev, u32 queues) argument
997 if (queues & BIT(IEEE80211_AC_VO))
999 if (queues & BIT(IEEE80211_AC_VI))
1001 if (queues & BIT(IEEE80211_AC_BE))
1003 if (queues & BIT(IEEE80211_AC_BK))
1056 void rtw_mac_flush_queues(struct rtw_dev *rtwdev, u32 queues, bool drop) argument
1060 /* If all of the hardware queues are requested to flush,
1061 * or the priority queues are not mapped yet,
1062 * flush all of the priority queues
1064 if (queues
[all...]
/linux-master/drivers/media/platform/qcom/venus/
H A Dhfi_venus.c127 struct iface_queue queues[IFACEQ_NUM]; member in struct:venus_hfi_device
396 queue = &hdev->queues[IFACEQ_CMD_IDX];
409 queue = &hdev->queues[IFACEQ_MSG_IDX];
687 queue = &hdev->queues[IFACEQ_MSG_IDX];
721 queue = &hdev->queues[IFACEQ_DBG_IDX];
770 memset(hdev->queues, 0, sizeof(hdev->queues));
795 queue = &hdev->queues[i];
827 queue = &hdev->queues[IFACEQ_DBG_IDX];
924 qhdr = hdev->queues[inde
[all...]
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_device.c128 * active queues and calls pvr_queue_process() on them.
135 mutex_lock(&pvr_dev->queues.lock);
137 /* Move all active queues to a temporary list. Queues that remain
139 * the queues.active list by pvr_queue_process().
141 list_splice_init(&pvr_dev->queues.active, &active_queues);
146 mutex_unlock(&pvr_dev->queues.lock);
/linux-master/drivers/scsi/aacraid/
H A Dcommsup.c343 * how these queues are implemented.
365 * All of the queues wrap when they reach the end, so we check
371 q = &dev->queues->queue[qid];
580 if (!dev->queues)
652 struct aac_queue * q = &dev->queues->queue[AdapNormCmdQueue];
920 q = &dev->queues->queue[AdapNormRespQueue];
1558 kfree(aac->queues);
1559 aac->queues = NULL;
2165 t_lock = dev->queues->queue[HostNormCmdQueue].lock;
2168 while (!list_empty(&(dev->queues
[all...]
H A Drx.c55 aac_command_normal(&dev->queues->queue[HostNormCmdQueue]);
59 aac_response_normal(&dev->queues->queue[HostNormRespQueue]);
400 struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
423 struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
/linux-master/net/mac80211/
H A Dutil.c397 if (local->hw.queues < IEEE80211_NUM_ACS)
400 for (i = 0; i < local->hw.queues; i++) {
442 if (WARN_ON(queue >= hw->queues))
507 if (WARN_ON(queue >= hw->queues))
586 for (i = 0; i < hw->queues; i++)
594 unsigned long queues,
604 for_each_set_bit(i, &queues, hw->queues)
624 if (WARN_ON(queue >= hw->queues))
636 unsigned long queues,
593 ieee80211_stop_queues_by_reason(struct ieee80211_hw *hw, unsigned long queues, enum queue_stop_reason reason, bool refcounted) argument
635 ieee80211_wake_queues_by_reason(struct ieee80211_hw *hw, unsigned long queues, enum queue_stop_reason reason, bool refcounted) argument
664 unsigned int queues; local
683 __ieee80211_flush_queues(struct ieee80211_local *local, struct ieee80211_sub_if_data *sdata, unsigned int queues, bool drop) argument
[all...]
H A Dwme.c129 local->hw.queues < IEEE80211_NUM_ACS)
H A Ddebugfs.c575 for (q = 0; q < local->hw.queues; q++)
585 DEBUGFS_READONLY_FILE_OPS(queues); variable
657 DEBUGFS_ADD(queues);
/linux-master/drivers/net/wireless/st/cw1200/
H A Dsta.h42 u32 queues, bool drop);
/linux-master/drivers/gpu/drm/msm/adreno/
H A Da6xx_gmu.h91 struct a6xx_hfi_queue queues[2]; member in struct:a6xx_gmu
/linux-master/drivers/block/null_blk/
H A Dnull_blk.h85 unsigned int submit_queues; /* number of submission queues */
86 unsigned int prev_submit_queues; /* number of submission queues before change */
87 unsigned int poll_queues; /* number of IOPOLL submission queues */
88 unsigned int prev_poll_queues; /* number of IOPOLL submission queues before change */
122 struct nullb_queue *queues; member in struct:nullb
/linux-master/drivers/net/wireless/mediatek/mt76/mt76x2/
H A Dpci_main.c109 u32 queues, bool drop)
108 mt76x2_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
/linux-master/tools/testing/selftests/net/forwarding/
H A Dethtool_mm.sh259 queues 1@0 1@1 1@2 1@3 \
273 queues 1@0 1@1 1@2 1@3 \
/linux-master/sound/virtio/
H A Dvirtio_card.c138 snd->queues[i].vqueue = vqs[i];
327 spin_lock_init(&snd->queues[i].lock);
/linux-master/drivers/target/
H A Dtarget_core_device.c693 dev->queues = kcalloc(nr_cpu_ids, sizeof(*dev->queues), GFP_KERNEL);
694 if (!dev->queues) {
703 q = &dev->queues[i];
1004 kfree(dev->queues);
/linux-master/net/sched/
H A Dsch_ets.c23 * When there is no traffic in any of the strict queues, the bandwidth-sharing
580 struct Qdisc *queues[TCQ_ETS_MAX_BANDS]; local
637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
640 if (!queues[i]) {
642 qdisc_put(queues[--i]);
668 q->classes[i].qdisc = queues[i];
/linux-master/drivers/misc/uacce/
H A Duacce.c175 list_add(&q->list, &uacce->queues);
532 INIT_LIST_HEAD(&uacce->queues);
591 list_for_each_entry_safe(q, next_q, &uacce->queues, list) {
608 /* disable sva now since no opened queues */

Completed in 239 milliseconds

123456789