/linux-master/drivers/net/wireless/ath/carl9170/ |
H A D | main.c | 211 while ((skb = __skb_dequeue(&tid_info->queue))) 264 while ((skb = __skb_dequeue(&tid_info->queue))) 328 #define CARL9170_FILL_QUEUE(queue, ai_fs, cwmin, cwmax, _txop) \ 330 queue.aifs = ai_fs; \ 331 queue.cw_min = cwmin; \ 332 queue.cw_max = cwmax; \ 333 queue.txop = _txop; \ 1369 unsigned int link_id, u16 queue, 1376 memcpy(&ar->edcf[ar9170_qmap(queue)], param, sizeof(*param)); 1429 skb_queue_head_init(&tid_info->queue); 1367 carl9170_op_conf_tx(struct ieee80211_hw *hw, struct ieee80211_vif *vif, unsigned int link_id, u16 queue, const struct ieee80211_tx_queue_params *param) argument [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | ops.c | 289 * and mvm->mutex. Will be handled with the wiphy_work queue infra 290 * instead of regular work queue. 1283 * If we have the new TX/queue allocation API initialize them 1286 * time (e.g. P2P Device is optional), and if a dynamic queue 1396 /* Set a short watchdog for the command queue */ 1804 static int iwl_mvm_is_static_queue(struct iwl_mvm *mvm, int queue) argument 1806 return queue == mvm->aux_queue || queue == mvm->probe_queue || 1807 queue == mvm->p2p_dev_queue || queue 2128 iwl_mvm_rx_mq_rss(struct iwl_op_mode *op_mode, struct napi_struct *napi, struct iwl_rx_cmd_buffer *rxb, unsigned int queue) argument [all...] |
H A D | mvm.h | 74 /* offchannel queue towards mac80211 */ 301 * @mgmt_queue: queue number for unbufferable management frames 304 * @cab_queue: content-after-beacon (multicast) queue 730 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer 734 * @queue: queue of this reorder buffer 737 * @valid: reordering is valid for this queue 745 int queue; member in struct:iwl_mvm_reorder_buffer 754 * struct iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno 772 * @entries_per_queue: # of buffers per queue, thi 1487 iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue) argument 1493 iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue) argument [all...] |
/linux-master/drivers/media/pci/intel/ipu3/ |
H A D | ipu3-cio2.c | 821 /* Initialize buffer queue */ 902 dev_dbg(dev, "queue buffer %d\n", vb->index); 908 * does not get an opportunity to queue new buffers to the CIO2 DMA 937 * hw, the user has requested N buffer queue. The driver 1380 if (cio2->queue[s_asd->csi2.port].sensor) 1387 q = &cio2->queue[s_asd->csi2.port]; 1404 cio2->queue[s_asd->csi2.port].sensor = NULL; 1418 q = &cio2->queue[s_asd->csi2.port]; 1564 CIO2_ENTITY_NAME " %td", q - cio2->queue); 1585 dev_err(dev, "failed to initialize videobuf2 queue ( [all...] |
/linux-master/drivers/dma/dw/ |
H A D | core.c | 69 list_add_tail(&desc->desc_node, &dwc->queue); 184 /* The tasklet will hopefully advance the queue... */ 220 if (list_empty(&dwc->queue)) 223 list_move(dwc->queue.next, &dwc->active_list); 437 list_move(dwc->queue.next, dwc->active_list.prev); 869 list_splice_init(&dwc->queue, &list); 1022 BUG_ON(!list_empty(&dwc->queue)); 1171 INIT_LIST_HEAD(&dwc->queue);
|
/linux-master/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_pf.c | 28 /* We assume here Tx and Rx queue 0 are owned by the PF */ 30 /* Shut off VF access to their queues forcing them to queue 0 */ 174 /* record maximum queue count, we limit ourselves to 128 */ 503 u16 vsi, queue, pc, q_idx; local 524 for (queue = 0; queue < queue_count; queue++, q_idx++) { 540 for (queue = 0; queue < queue_count; queue [all...] |
/linux-master/drivers/block/ |
H A D | sunvdc.c | 183 /* restart blk queue when ring is half emptied. also called after 188 blk_mq_start_stopped_hw_queues(port->disk->queue, true); 540 struct vdc_port *port = hctx->queue->queuedata; 844 q = g->queue; 852 g->queue = q; 1068 blk_mq_stop_hw_queues(port->disk->queue); 1114 struct request_queue *q = port->disk->queue; 1117 * Mark the queue as draining, then freeze/quiesce to ensure 1142 pr_warn(PFX "%s ldc down %llu seconds, draining queue\n", 1171 blk_mq_stop_hw_queues(port->disk->queue); [all...] |
/linux-master/net/phonet/ |
H A D | pep.c | 337 struct sk_buff_head *queue; local 383 queue = &pn->ctrlreq_queue; 384 goto queue; 405 queue = &sk->sk_receive_queue; 406 goto queue; 450 queue: 453 skb_queue_tail(queue, skb); 652 * We then queue the skb to the right connected sock (if any).
|
/linux-master/include/linux/ |
H A D | skbuff.h | 757 * @list: queue head 808 * @ooo_okay: allow the mapping of a socket to a queue to be changed 1821 * skb_queue_empty - check if a queue is empty 1822 * @list: queue head 1824 * Returns true if the queue is empty, false otherwise. 1832 * skb_queue_empty_lockless - check if a queue is empty 1833 * @list: queue head 1835 * Returns true if the queue is empty, false otherwise. 1845 * skb_queue_is_last - check if skb is the last entry in the queue 1846 * @list: queue hea [all...] |
/linux-master/drivers/usb/dwc2/ |
H A D | gadget.c | 404 INIT_LIST_HEAD(&req->queue); 572 "%s: no queue slots available (0x%08x)\n", 992 if (list_empty(&hs_ep->queue)) { 994 dev_dbg(hsotg->dev, "%s: No requests in queue\n", __func__); 1007 list_for_each_entry_safe(hs_req, treq, &hs_ep->queue, queue) { 1039 * dwc2_hsotg_start_req - start a USB request from an endpoint's queue 1257 * We've been asked to queue a request, so ensure that the memory buffer 1394 dev_err(hsotg->dev, "invalid EP 0 state in queue %d\n", 1430 INIT_LIST_HEAD(&hs_req->queue); [all...] |
/linux-master/drivers/scsi/ |
H A D | pmcraid.c | 158 list_for_each_entry(temp, &pinstance->used_res_q, queue) { 202 * @lim: queue limits 207 * and max_sectors value will be over-written to 512. It also sets queue depth 274 * pmcraid_change_queue_depth - Change the device's queue depth 1573 list_for_each_entry(res, &pinstance->used_res_q, queue) { 1609 struct pmcraid_resource_entry, queue); 1611 list_del(&res->queue); 1614 list_add_tail(&res->queue, &pinstance->used_res_q); 1632 list_move_tail(&res->queue, &pinstance->free_res_q); 2469 * This function sends a cancel all to a device to clear the queue [all...] |
/linux-master/drivers/nvme/host/ |
H A D | core.c | 142 * Only new queue scan work when admin and IO queues are both alive 711 } else { /* no queuedata implies admin queue */ 772 * Only allow commands on a live queue, except for the connect 773 * command, which is require to set the queue live in the 1645 * Degraded controllers might return an error when setting the queue 1647 * access to the admin queue, as that might be only way to fix them up. 1650 dev_err(ctrl->device, "Could not set queue count (%d)\n", status); 2029 if (blk_queue_is_zoned(ns->disk->queue)) { 2045 blk_mq_freeze_queue(ns->disk->queue); 2046 lim = queue_limits_start_update(ns->disk->queue); [all...] |
/linux-master/drivers/net/ethernet/synopsys/ |
H A D | dwc-xlgmac-hw.c | 515 /* Enable each Tx queue */ 552 /* Disable each Tx queue */ 574 unsigned int queue) 580 * packets. Wait for the Rx queue to empty the Rx fifo. Don't 585 rx_status = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_RQDR)); 598 "timed out waiting for Rx queue %u to empty\n", 599 queue); 619 /* Enable each Rx queue */ 660 /* Disable each Rx queue */ 1507 unsigned int qptc, qptc_extra, queue; local 573 xlgmac_prepare_rx_stop(struct xlgmac_pdata *pdata, unsigned int queue) argument [all...] |
/linux-master/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_ethtool.c | 311 /* Account for queue statistics */ 330 /* Account for non-queue statistics */ 1128 /* Reset the indirection table if rx queue count is updated */ 1568 /* The packet is expected to receive on rx-queue 0 even though RSS is 1569 * enabled. This is because the queue 0 is configured as the default 1570 * queue and that the loopback traffic is not IP. 2145 int qede_set_per_coalesce(struct net_device *dev, u32 queue, argument 2167 if (queue >= edev->num_queues) { 2168 DP_INFO(edev, "Invalid queue\n"); 2178 fp = &edev->fp_array[queue]; 2211 qede_get_per_coalesce(struct net_device *dev, u32 queue, struct ethtool_coalesce *coal) argument [all...] |
/linux-master/net/ipv6/ |
H A D | ip6_output.c | 1017 * Put this fragment into the sending queue. 1404 struct sk_buff_head *queue, 1433 skb = skb_peek_tail(queue); 1709 * Put the packet on the pending queue 1716 __skb_queue_tail(queue, skb); 1870 struct sk_buff_head *queue, 1884 skb = __skb_dequeue(queue); 1892 while ((tmp_skb = __skb_dequeue(queue)) != NULL) { 1980 struct sk_buff_head *queue, 1986 while ((skb = __skb_dequeue_tail(queue)) ! 1403 __ip6_append_data(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork_full, struct inet6_cork *v6_cork, struct page_frag *pfrag, int getfrag(void *from, char *to, int offset, int len, int odd, struct sk_buff *skb), void *from, size_t length, int transhdrlen, unsigned int flags, struct ipcm6_cookie *ipc6) argument 1869 __ip6_make_skb(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork, struct inet6_cork *v6_cork) argument 1979 __ip6_flush_pending_frames(struct sock *sk, struct sk_buff_head *queue, struct inet_cork_full *cork, struct inet6_cork *v6_cork) argument 2011 struct sk_buff_head queue; local [all...] |
/linux-master/drivers/net/wireless/st/cw1200/ |
H A D | wsm.c | 1058 arg.queue = (flags >> 1) & 3; 1457 struct cw1200_queue *queue) 1554 BUG_ON(cw1200_queue_remove(queue, wsm->packet_id)); 1627 /* Search for a queue with multicast frames buffered */ 1663 struct cw1200_queue *queue = NULL; local 1689 ret = wsm_get_tx_queue_and_mask(priv, &queue, 1691 queue_num = queue - priv->tx_queue; 1709 if (cw1200_queue_get(queue, 1715 tx_info, txpriv, queue)) 1730 (int)cw1200_queue_get_num_queued(queue, tx_allowed_mas 1453 wsm_handle_tx_data(struct cw1200_common *priv, struct wsm_tx *wsm, const struct ieee80211_tx_info *tx_info, const struct cw1200_txpriv *txpriv, struct cw1200_queue *queue) argument [all...] |
/linux-master/net/sunrpc/ |
H A D | cache.c | 293 * -ETIMEDOUT if upcall failed or request could not be queue or 388 INIT_LIST_HEAD(&cd->queue); 842 while (rp->q.list.next != &cd->queue && 848 if (rp->q.list.next == &cd->queue) { 983 for (cq= &rp->q; &cq->list != &cd->queue; 1009 for (cq= &rp->q; &cq->list != &cd->queue; 1040 list_add(&rp->q.list, &cd->queue); 1058 for (cq= &rp->q; &cq->list != &cd->queue; 1092 list_for_each_entry_safe(cq, tmp, &detail->queue, list) 1200 * register an upcall request to user-space and queue i [all...] |
/linux-master/drivers/media/platform/samsung/s5p-mfc/ |
H A D | s5p_mfc.c | 117 wake_up(&ctx->queue); 127 wake_up(&dev->queue); 606 wake_up(&ctx->queue); 615 wake_up(&ctx->queue); 642 wake_up(&ctx->queue); 759 wake_up(&ctx->queue); 786 init_waitqueue_head(&ctx->queue); 861 /* Init videobuf2 queue for CAPTURE */ 885 mfc_err("Failed to initialize videobuf2 queue(capture)\n"); 888 /* Init videobuf2 queue fo [all...] |
/linux-master/drivers/hsi/clients/ |
H A D | hsi_char.c | 75 * @rx_wait: RX requests wait queue 76 * @tx_wait: TX requests wait queue 117 struct list_head *queue) 122 list_add_tail(&msg->link, queue); 127 struct list_head *queue) 134 if (list_empty(queue)) 137 msg = list_first_entry(queue, struct hsi_msg, link); 116 hsc_add_tail(struct hsc_channel *channel, struct hsi_msg *msg, struct list_head *queue) argument 126 hsc_get_first_msg(struct hsc_channel *channel, struct list_head *queue) argument
|
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_spq.c | 223 /* QM physical queue */ 643 p_ent->queue = &p_spq->unlimited_pending; 648 p_ent->queue = &p_spq->pending; 692 if (p_ent->queue == &p_spq->unlimited_pending) { 723 /* entry is to be placed in 'pending' queue */ 875 /* Add the request to the pending queue */ 899 p_ent->queue == &p_spq->unlimited_pending); 901 if (p_ent->queue == &p_spq->unlimited_pending) {
|
/linux-master/drivers/net/wireless/ath/ath9k/ |
H A D | ar9003_hw.c | 1098 static bool ath9k_hw_verify_hang(struct ath_hw *ah, unsigned int queue) argument 1105 if (queue < 6) { 1107 reg_offset = queue * 5; 1110 reg_offset = (queue - 6) * 5; 1125 "MAC Hang signature found for queue: %d\n", queue); local
|
/linux-master/drivers/net/ethernet/sfc/ |
H A D | efx_channels.c | 140 * we allocate a TX queue per CPU. We share event queues across 156 * We need a channel per event queue, plus a VI per tx queue. 419 /* Create event queue 420 * Event queue memory allocations are done only once. If the channel 430 "chan %d create event queue\n", channel->channel); 432 /* Build an event queue with room for one event per tx and rx buffer, 442 /* Prepare channel's event queue */ 451 "chan %d init event queue\n", channel->channel); 462 /* Enable event queue processin [all...] |
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_dpmaif_rx.c | 88 dev_err(dpmaif_ctrl->dev, "RX queue %d has not been started\n", rxq->index); 525 /* Do not need to release if the queue is empty */ 1044 int t7xx_dpmaif_rxq_init(struct dpmaif_rx_queue *queue) argument 1048 ret = t7xx_dpmaif_rx_alloc(queue); 1050 dev_err(queue->dpmaif_ctrl->dev, "Failed to allocate RX buffers: %d\n", ret); 1055 void t7xx_dpmaif_rxq_free(struct dpmaif_rx_queue *queue) argument 1057 t7xx_dpmaif_rx_buf_free(queue); 1109 * Wait for all the RX work to finish executing and mark the RX queue as paused.
|
/linux-master/drivers/crypto/ |
H A D | img-hash.c | 133 struct crypto_queue queue; member in struct:img_hash_dev 511 res = ahash_enqueue_request(&hdev->queue, req); 518 backlog = crypto_get_backlog(&hdev->queue); 519 async_req = crypto_dequeue_request(&hdev->queue); 747 "Attempt to read from an empty result queue\n"); 959 crypto_init_queue(&hdev->queue, IMG_HASH_QUEUE_LENGTH);
|
/linux-master/drivers/rapidio/devices/ |
H A D | tsi721_dma.c | 546 * the pending queue. 548 if (!desc && !bdma_chan->active_tx && !list_empty(&bdma_chan->queue)) { 549 desc = list_first_entry(&bdma_chan->queue, 716 list_add_tail(&desc->desc_node, &bdma_chan->queue); 741 /* Allocate queue of transaction descriptors */ 924 list_splice_init(&bdma_chan->queue, &list); 990 INIT_LIST_HEAD(&bdma_chan->queue);
|