/linux-master/drivers/crypto/hisilicon/sec/ |
H A D | sec_drv.c | 639 if (!info->queues[i].in_use) { 640 info->queues[i].in_use = true; 644 return &info->queues[i]; 976 /* Same QoS for all queues */ 1236 ret = sec_queue_config(info, &info->queues[i], i); 1240 ret = sec_queue_irq_init(&info->queues[i]); 1242 sec_queue_unconfig(info, &info->queues[i]); 1266 sec_queue_irq_uninit(&info->queues[j]); 1267 sec_queue_unconfig(info, &info->queues[j]); 1285 sec_queue_irq_uninit(&info->queues[ [all...] |
/linux-master/drivers/scsi/arm/ |
H A D | acornscsi.c | 700 SCpnt = queue_remove_exclude(&host->queues.issue, host->busyluns); 708 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); 1746 if (!ok && queue_probetgtlun(&host->queues.disconnected, target, lun)) 1761 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); 1788 queue_add_cmd_tail(&host->queues.disconnected, host->SCpnt); 1798 host->SCpnt = queue_remove_tgtluntag(&host->queues.disconnected, 2407 * Purpose : queues a SCSI command 2435 host->stats.queues += 1; 2440 if (!queue_add_cmd_ordered(&host->queues.issue, SCpnt)) { 2467 if (queue_remove_cmd(&host->queues [all...] |
/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 183 struct nvme_tcp_queue *queues; member in struct:nvme_tcp_ctrl 213 return queue - queue->ctrl->queues; 520 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; 541 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; 551 struct nvme_tcp_queue *queue = &ctrl->queues[0]; 1345 struct nvme_tcp_queue *queue = &ctrl->queues[0]; 1355 async->queue = &ctrl->queues[0]; 1362 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; 1650 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; 1824 struct nvme_tcp_queue *queue = &ctrl->queues[qi [all...] |
H A D | fc.c | 155 struct nvme_fc_queue *queues; member in struct:nvme_fc_ctrl 1404 * queues) that are part of the association. E.g. things are torn 1944 * allocates a single tag set for all io queues and sizes 1945 * the io queues to fully hold all possible tags. Thus, the 2147 struct nvme_fc_queue *queue = &ctrl->queues[queue_idx]; 2180 ret = __nvme_fc_init_request(ctrl, &ctrl->queues[0], 2219 struct nvme_fc_queue *queue = &ctrl->queues[qidx]; 2244 queue = &ctrl->queues[idx]; 2309 nvme_fc_free_queue(&ctrl->queues[i]); 2329 struct nvme_fc_queue *queue = &ctrl->queues[ctr [all...] |
H A D | pci.c | 97 "Number of queues to use for writes. If not set, reads and writes " 102 MODULE_PARM_DESC(poll_queues, "Number of queues to use for polled IO."); 119 struct nvme_queue *queues; member in struct:nvme_dev 193 /* only used for poll queues: */ 342 nvme_dbbuf_free(&dev->queues[i]); 399 struct nvme_queue *nvmeq = &dev->queues[0]; 412 struct nvme_queue *nvmeq = &dev->queues[hctx_idx + 1]; 1133 struct nvme_queue *nvmeq = &dev->queues[0]; 1187 * set. Since URGENT priority is zeroes, it makes all queues 1430 nvme_free_queue(&dev->queues[ [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_debug.c | 51 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 160 list_for_each_entry(pqn, &pqm->queues, 222 list_for_each_entry(pqn, &pqm->queues, 327 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 336 list_for_each_entry(pqn, &pqm->queues, process_queue_list) 565 pr_debug("Resumed %d queues\n", resume_count); 624 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 1103 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
|
H A D | kfd_process_queue_manager.c | 37 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 71 pr_info("Cannot open more queues for process with pasid 0x%x\n", 166 INIT_LIST_HEAD(&pqm->queues); 209 list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) { 290 * On GFX 9.4.3, increase the number of queues that 306 * for debug process, verify that it is within the static queues limit 339 /* SDMA queues are always allocated statically no matter 420 list_add(&pqn->process_queue_list, &pqm->queues); 436 /* check if queues list is empty unregister process from device */ 643 list_for_each_entry(pqn, &pqm->queues, process_queue_lis [all...] |
H A D | kfd_packet_manager.c | 122 struct list_head *queues, 145 pr_debug("Building runlist ib process count: %d queues count %d\n", 149 list_for_each_entry(cur, queues, list) { 206 pr_debug("Finished map process and queues to runlist\n"); 121 pm_create_runlist_ib(struct packet_manager *pm, struct list_head *queues, uint64_t *rl_gpu_addr, size_t *rl_size_bytes) argument
|
/linux-master/drivers/net/wireless/realtek/rtw88/ |
H A D | mac.c | 992 static u32 get_priority_queues(struct rtw_dev *rtwdev, u32 queues) argument 997 if (queues & BIT(IEEE80211_AC_VO)) 999 if (queues & BIT(IEEE80211_AC_VI)) 1001 if (queues & BIT(IEEE80211_AC_BE)) 1003 if (queues & BIT(IEEE80211_AC_BK)) 1056 void rtw_mac_flush_queues(struct rtw_dev *rtwdev, u32 queues, bool drop) argument 1060 /* If all of the hardware queues are requested to flush, 1061 * or the priority queues are not mapped yet, 1062 * flush all of the priority queues 1064 if (queues [all...] |
/linux-master/drivers/media/platform/qcom/venus/ |
H A D | hfi_venus.c | 127 struct iface_queue queues[IFACEQ_NUM]; member in struct:venus_hfi_device 396 queue = &hdev->queues[IFACEQ_CMD_IDX]; 409 queue = &hdev->queues[IFACEQ_MSG_IDX]; 687 queue = &hdev->queues[IFACEQ_MSG_IDX]; 721 queue = &hdev->queues[IFACEQ_DBG_IDX]; 770 memset(hdev->queues, 0, sizeof(hdev->queues)); 795 queue = &hdev->queues[i]; 827 queue = &hdev->queues[IFACEQ_DBG_IDX]; 924 qhdr = hdev->queues[inde [all...] |
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_device.c | 128 * active queues and calls pvr_queue_process() on them. 135 mutex_lock(&pvr_dev->queues.lock); 137 /* Move all active queues to a temporary list. Queues that remain 139 * the queues.active list by pvr_queue_process(). 141 list_splice_init(&pvr_dev->queues.active, &active_queues); 146 mutex_unlock(&pvr_dev->queues.lock);
|
/linux-master/drivers/scsi/aacraid/ |
H A D | commsup.c | 343 * how these queues are implemented. 365 * All of the queues wrap when they reach the end, so we check 371 q = &dev->queues->queue[qid]; 580 if (!dev->queues) 652 struct aac_queue * q = &dev->queues->queue[AdapNormCmdQueue]; 920 q = &dev->queues->queue[AdapNormRespQueue]; 1558 kfree(aac->queues); 1559 aac->queues = NULL; 2165 t_lock = dev->queues->queue[HostNormCmdQueue].lock; 2168 while (!list_empty(&(dev->queues [all...] |
H A D | rx.c | 55 aac_command_normal(&dev->queues->queue[HostNormCmdQueue]); 59 aac_response_normal(&dev->queues->queue[HostNormRespQueue]); 400 struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue]; 423 struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
|
/linux-master/net/mac80211/ |
H A D | util.c | 397 if (local->hw.queues < IEEE80211_NUM_ACS) 400 for (i = 0; i < local->hw.queues; i++) { 442 if (WARN_ON(queue >= hw->queues)) 507 if (WARN_ON(queue >= hw->queues)) 586 for (i = 0; i < hw->queues; i++) 594 unsigned long queues, 604 for_each_set_bit(i, &queues, hw->queues) 624 if (WARN_ON(queue >= hw->queues)) 636 unsigned long queues, 593 ieee80211_stop_queues_by_reason(struct ieee80211_hw *hw, unsigned long queues, enum queue_stop_reason reason, bool refcounted) argument 635 ieee80211_wake_queues_by_reason(struct ieee80211_hw *hw, unsigned long queues, enum queue_stop_reason reason, bool refcounted) argument 664 unsigned int queues; local 683 __ieee80211_flush_queues(struct ieee80211_local *local, struct ieee80211_sub_if_data *sdata, unsigned int queues, bool drop) argument [all...] |
H A D | wme.c | 129 local->hw.queues < IEEE80211_NUM_ACS)
|
H A D | debugfs.c | 575 for (q = 0; q < local->hw.queues; q++) 585 DEBUGFS_READONLY_FILE_OPS(queues); variable 657 DEBUGFS_ADD(queues);
|
/linux-master/drivers/net/wireless/st/cw1200/ |
H A D | sta.h | 42 u32 queues, bool drop);
|
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a6xx_gmu.h | 91 struct a6xx_hfi_queue queues[2]; member in struct:a6xx_gmu
|
/linux-master/drivers/block/null_blk/ |
H A D | null_blk.h | 85 unsigned int submit_queues; /* number of submission queues */ 86 unsigned int prev_submit_queues; /* number of submission queues before change */ 87 unsigned int poll_queues; /* number of IOPOLL submission queues */ 88 unsigned int prev_poll_queues; /* number of IOPOLL submission queues before change */ 122 struct nullb_queue *queues; member in struct:nullb
|
/linux-master/drivers/net/wireless/mediatek/mt76/mt76x2/ |
H A D | pci_main.c | 109 u32 queues, bool drop) 108 mt76x2_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
|
/linux-master/tools/testing/selftests/net/forwarding/ |
H A D | ethtool_mm.sh | 259 queues 1@0 1@1 1@2 1@3 \ 273 queues 1@0 1@1 1@2 1@3 \
|
/linux-master/sound/virtio/ |
H A D | virtio_card.c | 138 snd->queues[i].vqueue = vqs[i]; 327 spin_lock_init(&snd->queues[i].lock);
|
/linux-master/drivers/target/ |
H A D | target_core_device.c | 693 dev->queues = kcalloc(nr_cpu_ids, sizeof(*dev->queues), GFP_KERNEL); 694 if (!dev->queues) { 703 q = &dev->queues[i]; 1004 kfree(dev->queues);
|
/linux-master/net/sched/ |
H A D | sch_ets.c | 23 * When there is no traffic in any of the strict queues, the bandwidth-sharing 580 struct Qdisc *queues[TCQ_ETS_MAX_BANDS]; local 637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 640 if (!queues[i]) { 642 qdisc_put(queues[--i]); 668 q->classes[i].qdisc = queues[i];
|
/linux-master/drivers/misc/uacce/ |
H A D | uacce.c | 175 list_add(&q->list, &uacce->queues); 532 INIT_LIST_HEAD(&uacce->queues); 591 list_for_each_entry_safe(q, next_q, &uacce->queues, list) { 608 /* disable sva now since no opened queues */
|