/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 183 struct nvme_tcp_queue *queues; member in struct:nvme_tcp_ctrl 213 return queue - queue->ctrl->queues; 520 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; 541 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; 551 struct nvme_tcp_queue *queue = &ctrl->queues[0]; 1345 struct nvme_tcp_queue *queue = &ctrl->queues[0]; 1355 async->queue = &ctrl->queues[0]; 1362 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; 1650 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; 1824 struct nvme_tcp_queue *queue = &ctrl->queues[qi [all...] |
H A D | pci.c | 97 "Number of queues to use for writes. If not set, reads and writes " 102 MODULE_PARM_DESC(poll_queues, "Number of queues to use for polled IO."); 119 struct nvme_queue *queues; member in struct:nvme_dev 193 /* only used for poll queues: */ 342 nvme_dbbuf_free(&dev->queues[i]); 399 struct nvme_queue *nvmeq = &dev->queues[0]; 412 struct nvme_queue *nvmeq = &dev->queues[hctx_idx + 1]; 1133 struct nvme_queue *nvmeq = &dev->queues[0]; 1187 * set. Since URGENT priority is zeroes, it makes all queues 1430 nvme_free_queue(&dev->queues[ [all...] |
H A D | fc.c | 155 struct nvme_fc_queue *queues; member in struct:nvme_fc_ctrl 1404 * queues) that are part of the association. E.g. things are torn 1944 * allocates a single tag set for all io queues and sizes 1945 * the io queues to fully hold all possible tags. Thus, the 2147 struct nvme_fc_queue *queue = &ctrl->queues[queue_idx]; 2180 ret = __nvme_fc_init_request(ctrl, &ctrl->queues[0], 2219 struct nvme_fc_queue *queue = &ctrl->queues[qidx]; 2244 queue = &ctrl->queues[idx]; 2309 nvme_fc_free_queue(&ctrl->queues[i]); 2329 struct nvme_fc_queue *queue = &ctrl->queues[ctr [all...] |
/linux-master/net/mac80211/ |
H A D | tx.c | 1684 if (WARN_ON_ONCE(q >= local->hw.queues)) { 1699 * Drop off-channel frames if queues 1803 * are dequeued from the intermediate queues (when they are enabled). 4780 for (i = 0; i < local->hw.queues; i++) { 4854 for (i = 0; i < local->hw.queues; i++) { 5965 u32 queues; local 6006 queues = BIT(sdata->vif.hw_queue[ieee802_1d_to_ac[tid]]); 6007 __ieee80211_flush_queues(local, sdata, queues, false); 6212 if (local->hw.queues < IEEE80211_NUM_ACS)
|
H A D | mlme.c | 2553 for (q = 0; q < local->hw.queues; q++) { 2621 if (local->hw.queues < IEEE80211_NUM_ACS) 2752 if (local->hw.queues < IEEE80211_NUM_ACS) 3576 * frame but queues will be stopped. 5259 local->hw.queues >= IEEE80211_NUM_ACS; 5488 /* get uapsd queues configuration - same for all links */ 7906 wmm_used = bss->wmm_used && (local->hw.queues >= IEEE80211_NUM_ACS); 8297 (local->hw.queues >= IEEE80211_NUM_ACS);
|
H A D | ieee80211_i.h | 72 * It makes little sense to enable u-APSD queues by default because it needs 98 /* Keep a station's queues on the active list for deficit accounting 539 * Bitmask of enabled u-apsd queues, 579 * while the sdata->work is flushing the queues, for example. 834 * mode, so queues are stopped 1432 * queues increases over the limit. */ 2356 unsigned long queues, 2366 unsigned long queues, 2383 unsigned int queues, bool drop);
|
H A D | cfg.c | 1944 local->hw.queues >= IEEE80211_NUM_ACS) 2779 if (local->hw.queues < IEEE80211_NUM_ACS) 4399 * queues.
|
/linux-master/include/net/ |
H A D | mac80211.h | 93 * driver to keep hardware queues short and to provide some fairness between 100 * Intermediate queues (struct ieee80211_txq) are kept per-sta per-tid, with 107 * The driver can't access the internal TX queues (iTXQs) directly. 114 * Drivers can optionally delegate responsibility for scheduling queues to 149 * enum ieee80211_max_queues - maximum number of queues 151 * @IEEE80211_MAX_QUEUES: Maximum number of regular device queues. 152 * @IEEE80211_MAX_QUEUE_MAP: bitmap with maximum queues set 840 * that a frame can be transmitted while the queues are stopped for 2411 * @uapsd_queues: bitmap of queues configured for uapsd. Only valid 2428 * @txq: per-TID data TX queues; not 2991 u16 queues; member in struct:ieee80211_hw [all...] |
/linux-master/drivers/net/wireless/virtual/ |
H A D | mac80211_hwsim.c | 924 /* MAC80211_HWSIM virtio queues */ 2887 u32 queues, bool drop) 2889 /* Not implemented, queues only on kernel side */ 5262 hw->queues = 5; 2885 mac80211_hwsim_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
|
/linux-master/drivers/net/wireless/ath/ath11k/ |
H A D | mac.c | 8157 u32 queues, bool drop) 10143 ar->hw->queues = ATH11K_HW_MAX_QUEUES; 8156 ath11k_mac_op_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci.c | 99 u8 count; /* number of queues in group */ 112 struct mlxsw_pci_queue_type_group queues[MLXSW_PCI_QUEUE_TYPE_COUNT]; member in struct:mlxsw_pci 187 return &mlxsw_pci->queues[q_type]; 213 return &mlxsw_pci->queues[q_type].q[q_num]; 1078 dev_err(&pdev->dev, "Unsupported number of queues\n"); 1097 dev_err(&pdev->dev, "Failed to initialize event queues\n"); 1104 dev_err(&pdev->dev, "Failed to initialize completion queues\n"); 1111 dev_err(&pdev->dev, "Failed to initialize send descriptor queues\n"); 1118 dev_err(&pdev->dev, "Failed to initialize receive descriptor queues\n"); 1122 /* We have to poll in command interface until queues ar [all...] |
/linux-master/net/core/ |
H A D | dev.c | 2307 * @txq: number of queues available 2324 netdev_warn(dev, "Number of in use tx queues changed invalidating tc mappings. Priority traffic classification disabled!\n"); 2335 netdev_warn(dev, "Number of in use tx queues changed. Priority %i to tc mapping %i is no longer valid. Setting map to 0\n", 2380 if (map->queues[pos] != index) 2384 map->queues[pos] = map->queues[--map->len]; 2487 if (map->queues[pos] != index) 2512 new_map->queues[i] = map->queues[i]; 2637 while ((pos < map->len) && (map->queues[po [all...] |
/linux-master/drivers/net/ |
H A D | tun.c | 113 /* MAX_TAP_QUEUES 256 is chosen to allow rx/tx queues to be equal 2798 int queues = ifr->ifr_flags & IFF_MULTI_QUEUE ? local 2823 NET_NAME_UNKNOWN, tun_setup, queues, 2824 queues);
|
H A D | xen-netfront.c | 67 "Maximum number of queues per virtual interface"); 167 struct netfront_queue *queues; member in struct:netfront_info 367 if (!np->queues || np->broken) 371 queue = &np->queues[i]; 660 queue = &np->queues[smp_processor_id() % num_queues]; 725 /* Drop the packet if no queues are set up */ 732 queue = &np->queues[queue_index]; 875 queue = &np->queues[i]; 886 struct netfront_queue *queue = &info->queues[i]; 893 kfree(info->queues); [all...] |
H A D | veth.c | 1308 pr_warn("Can't restore rx queues config %d -> %d %d", 1556 NL_SET_ERR_MSG_MOD(extack, "XDP expects number of rx queues not less than peer tx queues"); 1938 int queues = num_possible_cpus(); local 1940 if (queues > 4096) 1941 queues = 4096; 1942 return queues;
|
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a6xx_gpu_state.c | 1254 BUILD_BUG_ON(ARRAY_SIZE(gmu->queues) != ARRAY_SIZE(a6xx_state->hfi_queue_history)); 1256 for (i = 0; i < ARRAY_SIZE(gmu->queues); i++) { 1257 struct a6xx_hfi_queue *queue = &gmu->queues[i];
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 134 /* Mask out the reserved queues */ 294 list_for_each_entry(cur, &dqm->queues, list) { 364 /* For SDMA queues on SOC15 with 8-byte doorbell, use static 389 /* For CP queues on SOC15 */ 547 pr_warn("Can't create new usermode queue because %d queues were already created\n", 560 * Eviction state logic: mark all queues as evicted, even ones 561 * not currently active. Restoring inactive queues later only 634 pr_debug("Total of %d queues are accountable so far\n", 965 * called on multiple queues in a loop, so rather than locking/unlocking 1011 * called on multiple queues i [all...] |
H A D | kfd_chardev.c | 346 * on unmapped queues for usermode queue oversubscription (no aggregated doorbell) 1694 if (!list_empty(&p->pqm.queues)) { 1695 pr_debug("Process has user queues running\n"); 2066 /* Confirm all process queues are evicted */ 2068 pr_err("Cannot dump process when queues are not in evicted state\n"); 2561 * Set the process to evicted state to avoid running any new queues before all the memory 2615 pr_err("Failed to unpause queues ret:%d\n", ret);
|
H A D | kfd_priv.h | 85 * definitions for Kaveri. In Kaveri only the first ME queues participates 149 * Kernel module parameter to specify maximum number of supported queues per 191 * Don't evict process queues on vm fault 388 * @KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES: Preempts all queues in the 389 * running queues list. 391 * @KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES: Preempts all non-static queues 394 * @KFD_UNMAP_QUEUES_FILTER_BY_PASID: Preempts queues that belongs to 446 * @priority: Defines the queue priority relative to other queues in the 451 * currently all queues are initialized with the highest priority. 470 * @is_evicted: Defines if the queue is evicted. Only active queues 636 struct list_head queues; member in struct:process_queue_manager [all...] |
/linux-master/drivers/nvme/target/ |
H A D | fc.c | 170 struct nvmet_fc_tgt_queue *queues[NVMET_NR_QUEUES + 1]; member in struct:nvmet_fc_tgt_assoc 472 * queues) that are part of the association. E.g. things are torn 837 WARN_ON(assoc->queues[qid]); 838 assoc->queues[qid] = queue; 970 queue = assoc->queues[qid]; 1200 if (assoc->queues[i]) 1201 nvmet_fc_delete_target_queue(assoc->queues[i]); 1254 if (assoc->queues[i]) 1255 flush_workqueue(assoc->queues[i]->work_q); 1600 queue = assoc->queues[ [all...] |
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 90 MODULE_PARM_DESC(submit_queues, "Number of submission queues"); 94 MODULE_PARM_DESC(poll_queues, "Number of IOPOLL submission queues"); 186 MODULE_PARM_DESC(shared_tag_bitmap, "Use shared tag bitmap for all submission queues for blk-mq"); 214 MODULE_PARM_DESC(use_per_node_hctx, "Use per-node allocation for hardware context queues. Default: false"); 375 * Make sure that null_init_hctx() does not access nullb->queues[] past 1444 * number of hardware queues are prepared. If block layer failed 1445 * to prepare them, use previous numbers of submit queues and 1446 * poll queues to map queues. 1639 nq = &nullb->queues[hctx_id [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | mvm.h | 688 * aligned up to avoid cache line sharing between queues 897 struct work_struct add_stream_wk; /* To add streams to queues */ 1091 /* Tx queues */ 2235 /* Return a bitmask with all the hw supported queues, except for the 2688 u32 queues, bool drop);
|
H A D | mac80211.c | 408 * with many more queues than the current code can deal with 420 hw->queues = IEEE80211_NUM_ACS; 3228 * queues to this station will not be transmitted. The fw will 3900 * This is after the sync_rcu part, so the queues have already been 3902 * the queues. 5592 static void iwl_mvm_flush_no_vif(struct iwl_mvm *mvm, u32 queues, bool drop) argument 5600 iwl_mvm_flushable_queues(mvm) & queues); 5603 iwl_trans_wait_tx_queues_empty(mvm->trans, queues); 5627 u32 queues, bool drop) 5638 iwl_mvm_flush_no_vif(mvm, queues, dro 5626 iwl_mvm_mac_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument [all...] |
H A D | d3.c | 85 int tid, int queues) 93 /* and use the internal data for the other queues */ 94 for (i = 1; i < queues; i++) { 82 iwl_mvm_find_max_pn(struct ieee80211_key_conf *key, struct iwl_mvm_key_pn *ptk_pn, struct ieee80211_key_seq *seq, int tid, int queues) argument
|
/linux-master/include/linux/ |
H A D | netdevice.h | 742 * map is an array of queues. 748 u16 queues[]; member in struct:xps_map 751 #define XPS_MIN_MAP_ALLOC ((L1_CACHE_ALIGN(offsetof(struct xps_map, queues[1])) \ 1046 * transmit queues. 1147 * tx queues stopped. This allows the netdevice to perform queue 1858 * @queues_kset: Group of all Kobjects in the Tx and RX queues 1880 * @_rx: Array of RX queues 1881 * @num_rx_queues: Number of RX queues 1883 * @real_num_rx_queues: Number of RX queues currently active in device 1887 * allow to avoid NIC hard IRQ, on busy queues [all...] |