Searched refs:queues (Results 1 - 25 of 217) sorted by last modified time

123456789

/linux-master/drivers/nvme/host/
H A Dtcp.c183 struct nvme_tcp_queue *queues; member in struct:nvme_tcp_ctrl
213 return queue - queue->ctrl->queues;
520 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx];
541 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1];
551 struct nvme_tcp_queue *queue = &ctrl->queues[0];
1345 struct nvme_tcp_queue *queue = &ctrl->queues[0];
1355 async->queue = &ctrl->queues[0];
1362 struct nvme_tcp_queue *queue = &ctrl->queues[qid];
1650 struct nvme_tcp_queue *queue = &ctrl->queues[qid];
1824 struct nvme_tcp_queue *queue = &ctrl->queues[qi
[all...]
H A Dpci.c97 "Number of queues to use for writes. If not set, reads and writes "
102 MODULE_PARM_DESC(poll_queues, "Number of queues to use for polled IO.");
119 struct nvme_queue *queues; member in struct:nvme_dev
193 /* only used for poll queues: */
342 nvme_dbbuf_free(&dev->queues[i]);
399 struct nvme_queue *nvmeq = &dev->queues[0];
412 struct nvme_queue *nvmeq = &dev->queues[hctx_idx + 1];
1133 struct nvme_queue *nvmeq = &dev->queues[0];
1187 * set. Since URGENT priority is zeroes, it makes all queues
1430 nvme_free_queue(&dev->queues[
[all...]
H A Dfc.c155 struct nvme_fc_queue *queues; member in struct:nvme_fc_ctrl
1404 * queues) that are part of the association. E.g. things are torn
1944 * allocates a single tag set for all io queues and sizes
1945 * the io queues to fully hold all possible tags. Thus, the
2147 struct nvme_fc_queue *queue = &ctrl->queues[queue_idx];
2180 ret = __nvme_fc_init_request(ctrl, &ctrl->queues[0],
2219 struct nvme_fc_queue *queue = &ctrl->queues[qidx];
2244 queue = &ctrl->queues[idx];
2309 nvme_fc_free_queue(&ctrl->queues[i]);
2329 struct nvme_fc_queue *queue = &ctrl->queues[ctr
[all...]
/linux-master/net/mac80211/
H A Dtx.c1684 if (WARN_ON_ONCE(q >= local->hw.queues)) {
1699 * Drop off-channel frames if queues
1803 * are dequeued from the intermediate queues (when they are enabled).
4780 for (i = 0; i < local->hw.queues; i++) {
4854 for (i = 0; i < local->hw.queues; i++) {
5965 u32 queues; local
6006 queues = BIT(sdata->vif.hw_queue[ieee802_1d_to_ac[tid]]);
6007 __ieee80211_flush_queues(local, sdata, queues, false);
6212 if (local->hw.queues < IEEE80211_NUM_ACS)
H A Dmlme.c2553 for (q = 0; q < local->hw.queues; q++) {
2621 if (local->hw.queues < IEEE80211_NUM_ACS)
2752 if (local->hw.queues < IEEE80211_NUM_ACS)
3576 * frame but queues will be stopped.
5259 local->hw.queues >= IEEE80211_NUM_ACS;
5488 /* get uapsd queues configuration - same for all links */
7906 wmm_used = bss->wmm_used && (local->hw.queues >= IEEE80211_NUM_ACS);
8297 (local->hw.queues >= IEEE80211_NUM_ACS);
H A Dieee80211_i.h72 * It makes little sense to enable u-APSD queues by default because it needs
98 /* Keep a station's queues on the active list for deficit accounting
539 * Bitmask of enabled u-apsd queues,
579 * while the sdata->work is flushing the queues, for example.
834 * mode, so queues are stopped
1432 * queues increases over the limit. */
2356 unsigned long queues,
2366 unsigned long queues,
2383 unsigned int queues, bool drop);
H A Dcfg.c1944 local->hw.queues >= IEEE80211_NUM_ACS)
2779 if (local->hw.queues < IEEE80211_NUM_ACS)
4399 * queues.
/linux-master/include/net/
H A Dmac80211.h93 * driver to keep hardware queues short and to provide some fairness between
100 * Intermediate queues (struct ieee80211_txq) are kept per-sta per-tid, with
107 * The driver can't access the internal TX queues (iTXQs) directly.
114 * Drivers can optionally delegate responsibility for scheduling queues to
149 * enum ieee80211_max_queues - maximum number of queues
151 * @IEEE80211_MAX_QUEUES: Maximum number of regular device queues.
152 * @IEEE80211_MAX_QUEUE_MAP: bitmap with maximum queues set
840 * that a frame can be transmitted while the queues are stopped for
2411 * @uapsd_queues: bitmap of queues configured for uapsd. Only valid
2428 * @txq: per-TID data TX queues; not
2991 u16 queues; member in struct:ieee80211_hw
[all...]
/linux-master/drivers/net/wireless/virtual/
H A Dmac80211_hwsim.c924 /* MAC80211_HWSIM virtio queues */
2887 u32 queues, bool drop)
2889 /* Not implemented, queues only on kernel side */
5262 hw->queues = 5;
2885 mac80211_hwsim_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
/linux-master/drivers/net/wireless/ath/ath11k/
H A Dmac.c8157 u32 queues, bool drop)
10143 ar->hw->queues = ATH11K_HW_MAX_QUEUES;
8156 ath11k_mac_op_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
/linux-master/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci.c99 u8 count; /* number of queues in group */
112 struct mlxsw_pci_queue_type_group queues[MLXSW_PCI_QUEUE_TYPE_COUNT]; member in struct:mlxsw_pci
187 return &mlxsw_pci->queues[q_type];
213 return &mlxsw_pci->queues[q_type].q[q_num];
1078 dev_err(&pdev->dev, "Unsupported number of queues\n");
1097 dev_err(&pdev->dev, "Failed to initialize event queues\n");
1104 dev_err(&pdev->dev, "Failed to initialize completion queues\n");
1111 dev_err(&pdev->dev, "Failed to initialize send descriptor queues\n");
1118 dev_err(&pdev->dev, "Failed to initialize receive descriptor queues\n");
1122 /* We have to poll in command interface until queues ar
[all...]
/linux-master/net/core/
H A Ddev.c2307 * @txq: number of queues available
2324 netdev_warn(dev, "Number of in use tx queues changed invalidating tc mappings. Priority traffic classification disabled!\n");
2335 netdev_warn(dev, "Number of in use tx queues changed. Priority %i to tc mapping %i is no longer valid. Setting map to 0\n",
2380 if (map->queues[pos] != index)
2384 map->queues[pos] = map->queues[--map->len];
2487 if (map->queues[pos] != index)
2512 new_map->queues[i] = map->queues[i];
2637 while ((pos < map->len) && (map->queues[po
[all...]
/linux-master/drivers/net/
H A Dtun.c113 /* MAX_TAP_QUEUES 256 is chosen to allow rx/tx queues to be equal
2798 int queues = ifr->ifr_flags & IFF_MULTI_QUEUE ? local
2823 NET_NAME_UNKNOWN, tun_setup, queues,
2824 queues);
H A Dxen-netfront.c67 "Maximum number of queues per virtual interface");
167 struct netfront_queue *queues; member in struct:netfront_info
367 if (!np->queues || np->broken)
371 queue = &np->queues[i];
660 queue = &np->queues[smp_processor_id() % num_queues];
725 /* Drop the packet if no queues are set up */
732 queue = &np->queues[queue_index];
875 queue = &np->queues[i];
886 struct netfront_queue *queue = &info->queues[i];
893 kfree(info->queues);
[all...]
H A Dveth.c1308 pr_warn("Can't restore rx queues config %d -> %d %d",
1556 NL_SET_ERR_MSG_MOD(extack, "XDP expects number of rx queues not less than peer tx queues");
1938 int queues = num_possible_cpus(); local
1940 if (queues > 4096)
1941 queues = 4096;
1942 return queues;
/linux-master/drivers/gpu/drm/msm/adreno/
H A Da6xx_gpu_state.c1254 BUILD_BUG_ON(ARRAY_SIZE(gmu->queues) != ARRAY_SIZE(a6xx_state->hfi_queue_history));
1256 for (i = 0; i < ARRAY_SIZE(gmu->queues); i++) {
1257 struct a6xx_hfi_queue *queue = &gmu->queues[i];
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c134 /* Mask out the reserved queues */
294 list_for_each_entry(cur, &dqm->queues, list) {
364 /* For SDMA queues on SOC15 with 8-byte doorbell, use static
389 /* For CP queues on SOC15 */
547 pr_warn("Can't create new usermode queue because %d queues were already created\n",
560 * Eviction state logic: mark all queues as evicted, even ones
561 * not currently active. Restoring inactive queues later only
634 pr_debug("Total of %d queues are accountable so far\n",
965 * called on multiple queues in a loop, so rather than locking/unlocking
1011 * called on multiple queues i
[all...]
H A Dkfd_chardev.c346 * on unmapped queues for usermode queue oversubscription (no aggregated doorbell)
1694 if (!list_empty(&p->pqm.queues)) {
1695 pr_debug("Process has user queues running\n");
2066 /* Confirm all process queues are evicted */
2068 pr_err("Cannot dump process when queues are not in evicted state\n");
2561 * Set the process to evicted state to avoid running any new queues before all the memory
2615 pr_err("Failed to unpause queues ret:%d\n", ret);
H A Dkfd_priv.h85 * definitions for Kaveri. In Kaveri only the first ME queues participates
149 * Kernel module parameter to specify maximum number of supported queues per
191 * Don't evict process queues on vm fault
388 * @KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES: Preempts all queues in the
389 * running queues list.
391 * @KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES: Preempts all non-static queues
394 * @KFD_UNMAP_QUEUES_FILTER_BY_PASID: Preempts queues that belongs to
446 * @priority: Defines the queue priority relative to other queues in the
451 * currently all queues are initialized with the highest priority.
470 * @is_evicted: Defines if the queue is evicted. Only active queues
636 struct list_head queues; member in struct:process_queue_manager
[all...]
/linux-master/drivers/nvme/target/
H A Dfc.c170 struct nvmet_fc_tgt_queue *queues[NVMET_NR_QUEUES + 1]; member in struct:nvmet_fc_tgt_assoc
472 * queues) that are part of the association. E.g. things are torn
837 WARN_ON(assoc->queues[qid]);
838 assoc->queues[qid] = queue;
970 queue = assoc->queues[qid];
1200 if (assoc->queues[i])
1201 nvmet_fc_delete_target_queue(assoc->queues[i]);
1254 if (assoc->queues[i])
1255 flush_workqueue(assoc->queues[i]->work_q);
1600 queue = assoc->queues[
[all...]
/linux-master/drivers/block/null_blk/
H A Dmain.c90 MODULE_PARM_DESC(submit_queues, "Number of submission queues");
94 MODULE_PARM_DESC(poll_queues, "Number of IOPOLL submission queues");
186 MODULE_PARM_DESC(shared_tag_bitmap, "Use shared tag bitmap for all submission queues for blk-mq");
214 MODULE_PARM_DESC(use_per_node_hctx, "Use per-node allocation for hardware context queues. Default: false");
375 * Make sure that null_init_hctx() does not access nullb->queues[] past
1444 * number of hardware queues are prepared. If block layer failed
1445 * to prepare them, use previous numbers of submit queues and
1446 * poll queues to map queues.
1639 nq = &nullb->queues[hctx_id
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/
H A Dmvm.h688 * aligned up to avoid cache line sharing between queues
897 struct work_struct add_stream_wk; /* To add streams to queues */
1091 /* Tx queues */
2235 /* Return a bitmask with all the hw supported queues, except for the
2688 u32 queues, bool drop);
H A Dmac80211.c408 * with many more queues than the current code can deal with
420 hw->queues = IEEE80211_NUM_ACS;
3228 * queues to this station will not be transmitted. The fw will
3900 * This is after the sync_rcu part, so the queues have already been
3902 * the queues.
5592 static void iwl_mvm_flush_no_vif(struct iwl_mvm *mvm, u32 queues, bool drop) argument
5600 iwl_mvm_flushable_queues(mvm) & queues);
5603 iwl_trans_wait_tx_queues_empty(mvm->trans, queues);
5627 u32 queues, bool drop)
5638 iwl_mvm_flush_no_vif(mvm, queues, dro
5626 iwl_mvm_mac_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u32 queues, bool drop) argument
[all...]
H A Dd3.c85 int tid, int queues)
93 /* and use the internal data for the other queues */
94 for (i = 1; i < queues; i++) {
82 iwl_mvm_find_max_pn(struct ieee80211_key_conf *key, struct iwl_mvm_key_pn *ptk_pn, struct ieee80211_key_seq *seq, int tid, int queues) argument
/linux-master/include/linux/
H A Dnetdevice.h742 * map is an array of queues.
748 u16 queues[]; member in struct:xps_map
751 #define XPS_MIN_MAP_ALLOC ((L1_CACHE_ALIGN(offsetof(struct xps_map, queues[1])) \
1046 * transmit queues.
1147 * tx queues stopped. This allows the netdevice to perform queue
1858 * @queues_kset: Group of all Kobjects in the Tx and RX queues
1880 * @_rx: Array of RX queues
1881 * @num_rx_queues: Number of RX queues
1883 * @real_num_rx_queues: Number of RX queues currently active in device
1887 * allow to avoid NIC hard IRQ, on busy queues
[all...]

Completed in 531 milliseconds

123456789