Searched refs:queue (Results 726 - 750 of 1381) sorted by relevance

<<21222324252627282930>>

/linux-master/drivers/net/wireless/ath/carl9170/
H A Dmain.c211 while ((skb = __skb_dequeue(&tid_info->queue)))
264 while ((skb = __skb_dequeue(&tid_info->queue)))
328 #define CARL9170_FILL_QUEUE(queue, ai_fs, cwmin, cwmax, _txop) \
330 queue.aifs = ai_fs; \
331 queue.cw_min = cwmin; \
332 queue.cw_max = cwmax; \
333 queue.txop = _txop; \
1369 unsigned int link_id, u16 queue,
1376 memcpy(&ar->edcf[ar9170_qmap(queue)], param, sizeof(*param));
1429 skb_queue_head_init(&tid_info->queue);
1367 carl9170_op_conf_tx(struct ieee80211_hw *hw, struct ieee80211_vif *vif, unsigned int link_id, u16 queue, const struct ieee80211_tx_queue_params *param) argument
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/
H A Dops.c289 * and mvm->mutex. Will be handled with the wiphy_work queue infra
290 * instead of regular work queue.
1283 * If we have the new TX/queue allocation API initialize them
1286 * time (e.g. P2P Device is optional), and if a dynamic queue
1396 /* Set a short watchdog for the command queue */
1804 static int iwl_mvm_is_static_queue(struct iwl_mvm *mvm, int queue) argument
1806 return queue == mvm->aux_queue || queue == mvm->probe_queue ||
1807 queue == mvm->p2p_dev_queue || queue
2128 iwl_mvm_rx_mq_rss(struct iwl_op_mode *op_mode, struct napi_struct *napi, struct iwl_rx_cmd_buffer *rxb, unsigned int queue) argument
[all...]
H A Dmvm.h74 /* offchannel queue towards mac80211 */
301 * @mgmt_queue: queue number for unbufferable management frames
304 * @cab_queue: content-after-beacon (multicast) queue
730 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
734 * @queue: queue of this reorder buffer
737 * @valid: reordering is valid for this queue
745 int queue; member in struct:iwl_mvm_reorder_buffer
754 * struct iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
772 * @entries_per_queue: # of buffers per queue, thi
1487 iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue) argument
1493 iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue) argument
[all...]
/linux-master/drivers/media/pci/intel/ipu3/
H A Dipu3-cio2.c821 /* Initialize buffer queue */
902 dev_dbg(dev, "queue buffer %d\n", vb->index);
908 * does not get an opportunity to queue new buffers to the CIO2 DMA
937 * hw, the user has requested N buffer queue. The driver
1380 if (cio2->queue[s_asd->csi2.port].sensor)
1387 q = &cio2->queue[s_asd->csi2.port];
1404 cio2->queue[s_asd->csi2.port].sensor = NULL;
1418 q = &cio2->queue[s_asd->csi2.port];
1564 CIO2_ENTITY_NAME " %td", q - cio2->queue);
1585 dev_err(dev, "failed to initialize videobuf2 queue (
[all...]
/linux-master/drivers/dma/dw/
H A Dcore.c69 list_add_tail(&desc->desc_node, &dwc->queue);
184 /* The tasklet will hopefully advance the queue... */
220 if (list_empty(&dwc->queue))
223 list_move(dwc->queue.next, &dwc->active_list);
437 list_move(dwc->queue.next, dwc->active_list.prev);
869 list_splice_init(&dwc->queue, &list);
1022 BUG_ON(!list_empty(&dwc->queue));
1171 INIT_LIST_HEAD(&dwc->queue);
/linux-master/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_pf.c28 /* We assume here Tx and Rx queue 0 are owned by the PF */
30 /* Shut off VF access to their queues forcing them to queue 0 */
174 /* record maximum queue count, we limit ourselves to 128 */
503 u16 vsi, queue, pc, q_idx; local
524 for (queue = 0; queue < queue_count; queue++, q_idx++) {
540 for (queue = 0; queue < queue_count; queue
[all...]
/linux-master/drivers/block/
H A Dsunvdc.c183 /* restart blk queue when ring is half emptied. also called after
188 blk_mq_start_stopped_hw_queues(port->disk->queue, true);
540 struct vdc_port *port = hctx->queue->queuedata;
844 q = g->queue;
852 g->queue = q;
1068 blk_mq_stop_hw_queues(port->disk->queue);
1114 struct request_queue *q = port->disk->queue;
1117 * Mark the queue as draining, then freeze/quiesce to ensure
1142 pr_warn(PFX "%s ldc down %llu seconds, draining queue\n",
1171 blk_mq_stop_hw_queues(port->disk->queue);
[all...]
/linux-master/net/phonet/
H A Dpep.c337 struct sk_buff_head *queue; local
383 queue = &pn->ctrlreq_queue;
384 goto queue;
405 queue = &sk->sk_receive_queue;
406 goto queue;
450 queue:
453 skb_queue_tail(queue, skb);
652 * We then queue the skb to the right connected sock (if any).
/linux-master/include/linux/
H A Dskbuff.h757 * @list: queue head
808 * @ooo_okay: allow the mapping of a socket to a queue to be changed
1821 * skb_queue_empty - check if a queue is empty
1822 * @list: queue head
1824 * Returns true if the queue is empty, false otherwise.
1832 * skb_queue_empty_lockless - check if a queue is empty
1833 * @list: queue head
1835 * Returns true if the queue is empty, false otherwise.
1845 * skb_queue_is_last - check if skb is the last entry in the queue
1846 * @list: queue hea
[all...]
/linux-master/drivers/usb/dwc2/
H A Dgadget.c404 INIT_LIST_HEAD(&req->queue);
572 "%s: no queue slots available (0x%08x)\n",
992 if (list_empty(&hs_ep->queue)) {
994 dev_dbg(hsotg->dev, "%s: No requests in queue\n", __func__);
1007 list_for_each_entry_safe(hs_req, treq, &hs_ep->queue, queue) {
1039 * dwc2_hsotg_start_req - start a USB request from an endpoint's queue
1257 * We've been asked to queue a request, so ensure that the memory buffer
1394 dev_err(hsotg->dev, "invalid EP 0 state in queue %d\n",
1430 INIT_LIST_HEAD(&hs_req->queue);
[all...]
/linux-master/drivers/scsi/
H A Dpmcraid.c158 list_for_each_entry(temp, &pinstance->used_res_q, queue) {
202 * @lim: queue limits
207 * and max_sectors value will be over-written to 512. It also sets queue depth
274 * pmcraid_change_queue_depth - Change the device's queue depth
1573 list_for_each_entry(res, &pinstance->used_res_q, queue) {
1609 struct pmcraid_resource_entry, queue);
1611 list_del(&res->queue);
1614 list_add_tail(&res->queue, &pinstance->used_res_q);
1632 list_move_tail(&res->queue, &pinstance->free_res_q);
2469 * This function sends a cancel all to a device to clear the queue
[all...]
/linux-master/drivers/nvme/host/
H A Dcore.c142 * Only new queue scan work when admin and IO queues are both alive
711 } else { /* no queuedata implies admin queue */
772 * Only allow commands on a live queue, except for the connect
773 * command, which is require to set the queue live in the
1645 * Degraded controllers might return an error when setting the queue
1647 * access to the admin queue, as that might be only way to fix them up.
1650 dev_err(ctrl->device, "Could not set queue count (%d)\n", status);
2029 if (blk_queue_is_zoned(ns->disk->queue)) {
2045 blk_mq_freeze_queue(ns->disk->queue);
2046 lim = queue_limits_start_update(ns->disk->queue);
[all...]
/linux-master/drivers/net/ethernet/synopsys/
H A Ddwc-xlgmac-hw.c515 /* Enable each Tx queue */
552 /* Disable each Tx queue */
574 unsigned int queue)
580 * packets. Wait for the Rx queue to empty the Rx fifo. Don't
585 rx_status = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_RQDR));
598 "timed out waiting for Rx queue %u to empty\n",
599 queue);
619 /* Enable each Rx queue */
660 /* Disable each Rx queue */
1507 unsigned int qptc, qptc_extra, queue; local
573 xlgmac_prepare_rx_stop(struct xlgmac_pdata *pdata, unsigned int queue) argument
[all...]
/linux-master/drivers/net/ethernet/qlogic/qede/
H A Dqede_ethtool.c311 /* Account for queue statistics */
330 /* Account for non-queue statistics */
1128 /* Reset the indirection table if rx queue count is updated */
1568 /* The packet is expected to receive on rx-queue 0 even though RSS is
1569 * enabled. This is because the queue 0 is configured as the default
1570 * queue and that the loopback traffic is not IP.
2145 int qede_set_per_coalesce(struct net_device *dev, u32 queue, argument
2167 if (queue >= edev->num_queues) {
2168 DP_INFO(edev, "Invalid queue\n");
2178 fp = &edev->fp_array[queue];
2211 qede_get_per_coalesce(struct net_device *dev, u32 queue, struct ethtool_coalesce *coal) argument
[all...]
/linux-master/net/ipv6/
H A Dip6_output.c1017 * Put this fragment into the sending queue.
1404 struct sk_buff_head *queue,
1433 skb = skb_peek_tail(queue);
1709 * Put the packet on the pending queue
1716 __skb_queue_tail(queue, skb);
1870 struct sk_buff_head *queue,
1884 skb = __skb_dequeue(queue);
1892 while ((tmp_skb = __skb_dequeue(queue)) != NULL) {
1980 struct sk_buff_head *queue,
1986 while ((skb = __skb_dequeue_tail(queue)) !
1403 __ip6_append_data(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork_full, struct inet6_cork *v6_cork, struct page_frag *pfrag, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, size_t length, int transhdrlen, unsigned int flags, struct ipcm6_cookie *ipc6) argument
1869 __ip6_make_skb(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork, struct inet6_cork *v6_cork) argument
1979 __ip6_flush_pending_frames(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork, struct inet6_cork *v6_cork) argument
2011 struct sk_buff_head queue; local
[all...]
/linux-master/drivers/net/wireless/st/cw1200/
H A Dwsm.c1058 arg.queue = (flags >> 1) & 3;
1457 struct cw1200_queue *queue)
1554 BUG_ON(cw1200_queue_remove(queue, wsm->packet_id));
1627 /* Search for a queue with multicast frames buffered */
1663 struct cw1200_queue *queue = NULL; local
1689 ret = wsm_get_tx_queue_and_mask(priv, &queue,
1691 queue_num = queue - priv->tx_queue;
1709 if (cw1200_queue_get(queue,
1715 tx_info, txpriv, queue))
1730 (int)cw1200_queue_get_num_queued(queue, tx_allowed_mas
1453 wsm_handle_tx_data(struct cw1200_common *priv, struct wsm_tx *wsm, const struct ieee80211_tx_info *tx_info, const struct cw1200_txpriv *txpriv, struct cw1200_queue *queue) argument
[all...]
/linux-master/net/sunrpc/
H A Dcache.c293 * -ETIMEDOUT if upcall failed or request could not be queue or
388 INIT_LIST_HEAD(&cd->queue);
842 while (rp->q.list.next != &cd->queue &&
848 if (rp->q.list.next == &cd->queue) {
983 for (cq= &rp->q; &cq->list != &cd->queue;
1009 for (cq= &rp->q; &cq->list != &cd->queue;
1040 list_add(&rp->q.list, &cd->queue);
1058 for (cq= &rp->q; &cq->list != &cd->queue;
1092 list_for_each_entry_safe(cq, tmp, &detail->queue, list)
1200 * register an upcall request to user-space and queue i
[all...]
/linux-master/drivers/media/platform/samsung/s5p-mfc/
H A Ds5p_mfc.c117 wake_up(&ctx->queue);
127 wake_up(&dev->queue);
606 wake_up(&ctx->queue);
615 wake_up(&ctx->queue);
642 wake_up(&ctx->queue);
759 wake_up(&ctx->queue);
786 init_waitqueue_head(&ctx->queue);
861 /* Init videobuf2 queue for CAPTURE */
885 mfc_err("Failed to initialize videobuf2 queue(capture)\n");
888 /* Init videobuf2 queue fo
[all...]
/linux-master/drivers/hsi/clients/
H A Dhsi_char.c75 * @rx_wait: RX requests wait queue
76 * @tx_wait: TX requests wait queue
117 struct list_head *queue)
122 list_add_tail(&msg->link, queue);
127 struct list_head *queue)
134 if (list_empty(queue))
137 msg = list_first_entry(queue, struct hsi_msg, link);
116 hsc_add_tail(struct hsc_channel *channel, struct hsi_msg *msg, struct list_head *queue) argument
126 hsc_get_first_msg(struct hsc_channel *channel, struct list_head *queue) argument
/linux-master/drivers/net/ethernet/qlogic/qed/
H A Dqed_spq.c223 /* QM physical queue */
643 p_ent->queue = &p_spq->unlimited_pending;
648 p_ent->queue = &p_spq->pending;
692 if (p_ent->queue == &p_spq->unlimited_pending) {
723 /* entry is to be placed in 'pending' queue */
875 /* Add the request to the pending queue */
899 p_ent->queue == &p_spq->unlimited_pending);
901 if (p_ent->queue == &p_spq->unlimited_pending) {
/linux-master/drivers/net/wireless/ath/ath9k/
H A Dar9003_hw.c1098 static bool ath9k_hw_verify_hang(struct ath_hw *ah, unsigned int queue) argument
1105 if (queue < 6) {
1107 reg_offset = queue * 5;
1110 reg_offset = (queue - 6) * 5;
1125 "MAC Hang signature found for queue: %d\n", queue); local
/linux-master/drivers/net/ethernet/sfc/
H A Defx_channels.c140 * we allocate a TX queue per CPU. We share event queues across
156 * We need a channel per event queue, plus a VI per tx queue.
419 /* Create event queue
420 * Event queue memory allocations are done only once. If the channel
430 "chan %d create event queue\n", channel->channel);
432 /* Build an event queue with room for one event per tx and rx buffer,
442 /* Prepare channel's event queue */
451 "chan %d init event queue\n", channel->channel);
462 /* Enable event queue processin
[all...]
/linux-master/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif_rx.c88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index);
525 /* Do not need to release if the queue is empty */
1044 int t7xx_dpmaif_rxq_init(struct dpmaif_rx_queue *queue) argument
1048 ret = t7xx_dpmaif_rx_alloc(queue);
1050 dev_err(queue->dpmaif_ctrl->dev, "Failed to allocate RX buffers: %d\n", ret);
1055 void t7xx_dpmaif_rxq_free(struct dpmaif_rx_queue *queue) argument
1057 t7xx_dpmaif_rx_buf_free(queue);
1109 * Wait for all the RX work to finish executing and mark the RX queue as paused.
/linux-master/drivers/crypto/
H A Dimg-hash.c133 struct crypto_queue queue; member in struct:img_hash_dev
511 res = ahash_enqueue_request(&hdev->queue, req);
518 backlog = crypto_get_backlog(&hdev->queue);
519 async_req = crypto_dequeue_request(&hdev->queue);
747 "Attempt to read from an empty result queue\n");
959 crypto_init_queue(&hdev->queue, IMG_HASH_QUEUE_LENGTH);
/linux-master/drivers/rapidio/devices/
H A Dtsi721_dma.c546 * the pending queue.
548 if (!desc && !bdma_chan->active_tx && !list_empty(&bdma_chan->queue)) {
549 desc = list_first_entry(&bdma_chan->queue,
716 list_add_tail(&desc->desc_node, &bdma_chan->queue);
741 /* Allocate queue of transaction descriptors */
924 list_splice_init(&bdma_chan->queue, &list);
990 INIT_LIST_HEAD(&bdma_chan->queue);

Completed in 340 milliseconds

<<21222324252627282930>>