Searched refs:queue (Results 476 - 500 of 1381) sorted by relevance

<<11121314151617181920>>

/linux-master/drivers/usb/gadget/udc/
H A Ddummy_hcd.c85 struct list_head queue; member in struct:dummy_ep
98 struct list_head queue; /* ep's requests */ member in struct:dummy_request
333 while (!list_empty(&ep->queue)) {
336 req = list_entry(ep->queue.next, struct dummy_request, queue);
337 list_del_init(&req->queue);
670 INIT_LIST_HEAD(&req->queue);
684 WARN_ON(!list_empty(&req->queue));
702 if (!_req || !list_empty(&req->queue) || !_req->complete)
715 dev_dbg(udc_dev(dum), "ep %p queue re
[all...]
H A Dfsl_qe_udc.c82 /* the req->queue pointer is used by ep_queue() func, in which
83 * the request will be added into a udc_ep->queue 'd tail
84 * so here the req will be dropped from the ep->queue
86 list_del_init(&req->queue);
114 /* don't modify queue heads during completion callback */
131 while (!list_empty(&ep->queue)) {
133 req = list_entry(ep->queue.next, struct qe_req, queue);
780 if (ep->has_data <= 0 && (!list_empty(&ep->queue)))
907 if (list_empty(&ep->queue)) {
[all...]
H A Dmv_u3d_core.c180 dev_dbg(u3d->dev, "mv_u3d_done: remove req->queue\n");
181 /* Removed the req from ep queue */
182 list_del_init(&req->queue);
235 if (!list_empty(&ep->queue)) {
236 dev_err(u3d->dev, "add trb to non-empty queue!\n");
505 /* and push them to device queue */
510 /* irq handler advances the queue */
511 list_add_tail(&req->queue, &ep->queue);
674 INIT_LIST_HEAD(&req->queue);
[all...]
/linux-master/drivers/scsi/
H A Dips.c62 /* - Send all of the commands on the queue at once rather than */
103 /* - Don't release HA Lock in ips_next() until SC taken off queue */
798 /* See if the command is on the copp queue */
808 /* See if the command is on the wait queue */
1173 /* Set queue depths on devices once scan is complete */
2503 /* Take the next command off the queue and send it to the controller */
2522 * Block access to the queue function so
2627 spin_unlock(host->host_lock); /* Unlock HA after command is taken off queue */
2732 /* Add an item to the head of the queue */
2738 ips_putq_scb_head(ips_scb_queue_t * queue, ips_scb_ argument
2766 ips_removeq_scb_head(ips_scb_queue_t * queue) argument
2801 ips_removeq_scb(ips_scb_queue_t * queue, ips_scb_t * item) argument
2846 ips_putq_wait_tail(ips_wait_queue_entry_t *queue, struct scsi_cmnd *item) argument
2877 ips_removeq_wait_head(ips_wait_queue_entry_t *queue) argument
2911 ips_removeq_wait(ips_wait_queue_entry_t *queue, struct scsi_cmnd *item) argument
2958 ips_putq_copp_tail(ips_copp_queue_t * queue, ips_copp_wait_item_t * item) argument
2990 ips_removeq_copp_head(ips_copp_queue_t * queue) argument
3025 ips_removeq_copp(ips_copp_queue_t * queue, ips_copp_wait_item_t * item) argument
[all...]
/linux-master/block/
H A Dgenhd.c402 if (queue_is_mq(disk->queue) && disk->fops->poll_bio)
406 * The disk queue should now be all set with enough information about
408 * elevator if one is needed, that is, for devices requesting queue
411 elevator_init_mq(disk->queue);
537 rq_qos_exit(disk->queue);
552 if (disk->queue->elevator)
553 elevator_exit(disk->queue);
593 blk_queue_flag_set(QUEUE_FLAG_DYING, disk->queue);
603 blk_queue_start_drain(disk->queue);
641 struct request_queue *q = disk->queue;
[all...]
H A Dblk-cgroup.c162 * Both ->pd_free_fn() and request queue's release handler may
195 * example, don't try to follow throtl_data and request queue links.
311 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node);
319 if (!blk_get_queue(disk->queue))
322 blkg->q = disk->queue;
341 if (!blkcg_policy_enabled(disk->queue, pol))
360 blk_put_queue(disk->queue);
380 lockdep_assert_held(&disk->queue->queue_lock);
383 if (blk_queue_dying(disk->queue)) {
406 blkg->parent = blkg_lookup(blkcg_parent(blkcg), disk->queue);
[all...]
H A Dblk-zoned.c207 need_reset = blk_alloc_zone_bitmap(disk->queue->node, disk->nr_zones);
616 zwplug->wp_offset = sector & (disk->queue->limits.chunk_sectors - 1);
640 struct request_queue *q = zwplug->disk->queue;
822 sector += disk->queue->limits.chunk_sectors) {
837 * Grab an extra reference on the BIO request queue usage counter.
842 percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter);
956 * Drop the extra reference on the queue usage we got when
1324 * blk-mq devices will reuse the extra reference on the request queue
1330 blk_queue_exit(bdev->bd_disk->queue);
1402 * plugged BIOs to complete if there is a queue freez
[all...]
/linux-master/drivers/usb/gadget/udc/bdc/
H A Dbdc_ep.c530 list_add_tail(&req->queue, &ep->queue);
547 list_del(&req->queue);
574 /* de-queue any pending requests */
575 while (!list_empty(&ep->queue)) {
576 req = list_entry(ep->queue.next, struct bdc_req,
577 queue);
828 first_req = list_first_entry(&ep->queue, struct bdc_req,
829 queue);
974 if (unlikely(list_empty(&ep->queue))) {
[all...]
/linux-master/drivers/md/
H A Ddm-cache-policy-smq.c59 * from the queue itself.
246 * The stochastic-multi-queue is a set of lru lists stacked into levels.
253 struct queue { struct
270 static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels)
288 static unsigned int q_size(struct queue *q)
296 static void q_push(struct queue *q, struct entry *e)
306 static void q_push_front(struct queue *q, struct entry *e)
316 static void q_push_before(struct queue *q, struct entry *old, struct entry *e)
326 static void q_del(struct queue *q, struct entry *e)
336 static struct entry *q_peek(struct queue *
[all...]
/linux-master/drivers/net/wireless/intel/iwlwifi/queue/
H A Dtx.c14 #include "queue/tx.h"
181 * If skb is not NULL, it means that the whole queue is being
182 * freed and that the queue is not empty - free the skb
701 * should always be less than max_tfd_queue_size elements in the queue.
703 * to reserve any queue entries for this purpose.
733 "queue %d out of range", txq_id))
737 "TX on unused queue %d\n", txq_id))
774 /* Set up first empty entry in queue's array of Tx/cmd buffers */
800 /* start timer if queue currently empty */
848 /* just in case - this queue ma
1294 iwl_txq_dyn_free(struct iwl_trans *trans, int queue) argument
1308 "queue %d not used", queue); local
1334 struct iwl_txq *queue; local
1720 int queue; local
[all...]
/linux-master/drivers/atm/
H A Dnicstar.c228 while ((hb = skb_dequeue(&card->hbpool.queue)) != NULL) {
236 while ((iovb = skb_dequeue(&card->iovpool.queue)) != NULL) {
241 while ((lb = skb_dequeue(&card->lbpool.queue)) != NULL)
243 while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL)
655 skb_queue_head_init(&card->hbpool.queue);
669 skb_queue_tail(&card->hbpool.queue, hb);
674 skb_queue_head_init(&card->lbpool.queue);
688 skb_queue_tail(&card->lbpool.queue, lb);
710 skb_queue_head_init(&card->sbpool.queue);
724 skb_queue_tail(&card->sbpool.queue, s
[all...]
/linux-master/ipc/
H A Dsem.c129 /* One queue for each sleeping process in the system. */
131 struct list_head list; /* queue of pending operations */
223 * 2) queue.status: (SEM_BARRIER_2)
305 * merge_queues - merge single semop queues into global queue
308 * This function merges all per-semaphore queues into the global queue.
810 * update_queue is O(N^2) when it restarts scanning the whole queue of
833 * in the queue, then they have observed the original
844 * @wake_q: lockless wake-queue head.
871 /* operation completed, remove from queue & wakeup */
887 * @wake_q: lockless wake-queue hea
1993 struct sem_queue queue; local
[all...]
/linux-master/drivers/spi/
H A Dspi-topcliff-pch.c128 * @work: Reference to work queue handler
129 * @wait: Wait queue for waking up upon receiving an
134 * @queue: SPI Message queue
170 struct list_head queue; member in struct:pch_spi_data
482 /* add message to queue */
483 list_add_tail(&pmsg->queue, &data->queue);
489 dev_dbg(&pspi->dev, "%s - Invoked queue work\n", __func__);
562 /* flush queue an
[all...]
/linux-master/drivers/net/ethernet/huawei/hinic/
H A Dhinic_ethtool.c632 struct ethtool_coalesce *coal, u16 queue)
638 if (queue == COALESCE_ALL_QUEUE) {
643 if (queue >= nic_dev->num_qps) {
645 "Invalid queue_id: %d\n", queue);
648 rx_intr_coal_info = &nic_dev->rx_intr_coalesce[queue];
649 tx_intr_coal_info = &nic_dev->tx_intr_coalesce[queue];
710 "Failed to set %s queue%d coalesce",
718 u16 queue, bool set_rx_coal)
723 if (queue == COALESCE_ALL_QUEUE) {
731 if (queue >
631 __hinic_get_coalesce(struct net_device *netdev, struct ethtool_coalesce *coal, u16 queue) argument
716 __set_hw_coal_param(struct hinic_dev *nic_dev, struct hinic_intr_coal_info *intr_coal, u16 queue, bool set_rx_coal) argument
745 __hinic_set_coalesce(struct net_device *netdev, struct ethtool_coalesce *coal, u16 queue) argument
814 hinic_get_per_queue_coalesce(struct net_device *netdev, u32 queue, struct ethtool_coalesce *coal) argument
820 hinic_set_per_queue_coalesce(struct net_device *netdev, u32 queue, struct ethtool_coalesce *coal) argument
[all...]
/linux-master/drivers/scsi/bfa/
H A Dbfa_core.c700 * Callback only as long as there is room in request queue
766 int queue; local
773 * RME completion queue interrupt
777 for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++)
778 bfa_isr_rspq(bfa, queue);
786 * CPE completion queue interrupt
790 for (queue = 0; queue < BFI_IOC_MAX_CQ
804 int queue; local
[all...]
/linux-master/drivers/media/platform/rockchip/rkisp1/
H A Drkisp1-capture.c742 if (!list_empty(&cap->buf.queue)) {
745 cap->buf.next = list_first_entry(&cap->buf.queue, struct rkisp1_buffer, queue);
746 list_del(&cap->buf.next->queue);
867 static int rkisp1_vb2_queue_setup(struct vb2_queue *queue, argument
873 struct rkisp1_capture *cap = queue->drv_priv;
934 list_add_tail(&ispbuf->queue, &cap->buf.queue);
972 while (!list_empty(&cap->buf.queue)) {
973 buf = list_first_entry(&cap->buf.queue,
1123 rkisp1_vb2_stop_streaming(struct vb2_queue *queue) argument
1149 rkisp1_vb2_start_streaming(struct vb2_queue *queue, unsigned int count) argument
[all...]
/linux-master/drivers/s390/crypto/
H A Dap_bus.c78 /* Hashtable of all queue devices on the AP bus */
320 * ap_queue_info(): Check and get AP queue info.
324 * -1 invalid APQN, TAPQ error or AP queue status which
786 int rc, card, queue, devres, drvres; local
790 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid);
793 test_bit_inv(queue, ap_perms.aqm);
798 pr_debug("%s reprobing queue=%02x.%04x\n",
799 __func__, card, queue);
802 AP_DBF_WARN("%s reprobing queue=%02x.%04x failed\n",
803 __func__, card, queue);
826 ap_owned_by_def_drv(int card, int queue) argument
856 int card, queue, rc = 0; local
874 int card, queue, devres, drvres, rc = -ENODEV; local
[all...]
/linux-master/drivers/net/wireless/ath/ath5k/
H A Dath5k.h138 /* Access QCU registers per queue */
534 * @AR5K_TX_QUEUE_DATA: A normal data queue
535 * @AR5K_TX_QUEUE_BEACON: The beacon queue
536 * @AR5K_TX_QUEUE_CAB: The after-beacon queue
537 * @AR5K_TX_QUEUE_UAPSD: Unscheduled Automatic Power Save Delivery queue
571 * @AR5K_TX_QUEUE_ID_NOQCU_DATA: Data queue on AR5210 (no QCU available)
572 * @AR5K_TX_QUEUE_ID_NOQCU_BEACON: Beacon queue on AR5210 (no QCU available)
573 * @AR5K_TX_QUEUE_ID_DATA_MIN: Data queue min index
574 * @AR5K_TX_QUEUE_ID_DATA_MAX: Data queue max index
575 * @AR5K_TX_QUEUE_ID_CAB: Content after beacon queue
[all...]
/linux-master/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-dev.c463 unsigned int queue)
468 /* Does this queue handle the priority? */
469 if (pdata->prio2q_map[prio] != queue)
531 /* From MAC ver 30H the TFCR is per priority, instead of per queue */
2331 unsigned int queue,
2339 if (pdata->pfcq[queue] && (q_fifo_size > pdata->pfc_rfa)) {
2340 /* PFC is active for this queue */
2355 pdata->rx_rfa[queue] = 0;
2356 pdata->rx_rfd[queue] = 0;
2362 pdata->rx_rfa[queue]
462 xgbe_is_pfc_queue(struct xgbe_prv_data *pdata, unsigned int queue) argument
2330 xgbe_queue_flow_control_threshold(struct xgbe_prv_data *pdata, unsigned int queue, unsigned int q_fifo_size) argument
2659 unsigned int qptc, qptc_extra, queue; local
2740 unsigned int offset, queue, prio; local
3270 xgbe_txq_prepare_tx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument
3296 xgbe_prepare_tx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument
3384 xgbe_prepare_rx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_priv.h138 * eventually HQD registers when the queue is mapped by HWS. In order to
320 * page used by kernel queue
386 * enum kfd_unmap_queues_filter - Enum for queue filters.
405 * enum kfd_queue_type - Enum for various queue types.
407 * @KFD_QUEUE_TYPE_COMPUTE: Regular user mode queue type.
409 * @KFD_QUEUE_TYPE_SDMA: SDMA user mode queue type.
411 * @KFD_QUEUE_TYPE_HIQ: HIQ queue type.
413 * @KFD_QUEUE_TYPE_DIQ: DIQ queue type.
415 * @KFD_QUEUE_TYPE_SDMA_XGMI: Special SDMA queue for XGMI interface.
438 * @type: The queue typ
581 struct queue { struct
590 uint32_t queue; member in struct:queue
[all...]
/linux-master/drivers/s390/net/
H A Dqeth_core.h186 /* QDIO queue and buffer handling */
529 static inline void qeth_tx_arm_timer(struct qeth_qdio_out_q *queue, argument
532 timer_reduce(&queue->timer, usecs_to_jiffies(usecs) + jiffies);
535 static inline bool qeth_out_queue_is_full(struct qeth_qdio_out_q *queue) argument
537 return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q;
540 static inline bool qeth_out_queue_is_empty(struct qeth_qdio_out_q *queue) argument
542 return atomic_read(&queue->used_buffers) == 0;
913 struct qeth_qdio_out_q *queue)
915 return qeth_iqd_translate_txq(card->dev, queue->queue_no) ==
1103 struct qeth_qdio_out_q *queue, __be1
912 qeth_iqd_is_mcast_queue(struct qeth_card *card, struct qeth_qdio_out_q *queue) argument
[all...]
/linux-master/drivers/staging/media/ipu3/
H A Dipu3.c90 * Do not enable dummy buffers for master queue,
115 /* Allocate a dummy buffer for each queue where buffer is optional */
154 int queue, unsigned int pipe)
160 if (queue == IPU3_CSS_QUEUE_IN)
163 if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr))
168 if (imgu_css_buf_state(&imgu_pipe->queues[queue].dummybufs[i]) !=
175 imgu_css_buf_init(&imgu_pipe->queues[queue].dummybufs[i], queue,
176 imgu_pipe->queues[queue].dmap.daddr);
178 return &imgu_pipe->queues[queue]
153 imgu_dummybufs_get(struct imgu_device *imgu, int queue, unsigned int pipe) argument
[all...]
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_req.c45 struct rxe_queue *q = qp->sq.queue;
58 wqe = queue_addr_from_index(qp->sq.queue, wqe_index);
108 /* request a send queue retry */
126 q = qp->sq.queue;
164 struct rxe_queue *q = qp->sq.queue;
200 * @qp: the queue pair
209 * Requires ALL previous operations on the send queue
213 return qp->req.wqe_index != queue_get_consumer(qp->sq.queue,
583 qp->req.wqe_index = queue_next_index(qp->sq.queue,
634 qp->req.wqe_index = queue_next_index(qp->sq.queue, q
[all...]
/linux-master/fs/ocfs2/dlm/
H A Ddlmrecovery.c214 * 5) the new master collects up all of secondary lock queue info
217 * 6) each secondary lock queue responds with the full known lock info
220 * 8) upon receiving this message, the secondary queue node unlocks
849 /* queue up work for dlm_request_all_locks_worker */
1083 struct list_head *iter, *queue = &res->granted; local
1087 list_for_each(iter, queue)
1089 queue++;
1173 int queue)
1179 if (queue == DLM_BLOCKED_LIST)
1208 struct dlm_migratable_lockres *mres, int queue)
1171 dlm_prepare_lvb_for_migration(struct dlm_lock *lock, struct dlm_migratable_lockres *mres, int queue) argument
1207 dlm_add_lock_to_array(struct dlm_lock *lock, struct dlm_migratable_lockres *mres, int queue) argument
1264 struct list_head *queue; local
1792 struct list_head *queue, *iter; local
2048 struct list_head *queue; local
2212 struct list_head *queue; local
[all...]
/linux-master/drivers/block/
H A Dvirtio_blk.c134 struct virtio_blk *vblk = hctx->queue->queuedata;
230 return blk_rq_map_sg(hctx->queue, req, vbr->sg_table.sgl);
339 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata;
372 /* In case queue is stopped waiting for more buffers. */
374 blk_mq_start_stopped_hw_queues(vblk->disk->queue, true);
380 struct virtio_blk *vblk = hctx->queue->queuedata;
430 struct virtio_blk *vblk = hctx->queue->queuedata;
447 /* Don't stop the queue if -ENOMEM: we may have failed to
468 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata;
537 struct request_queue *q = vblk->disk->queue;
[all...]

Completed in 252 milliseconds

<<11121314151617181920>>