/linux-master/drivers/usb/gadget/udc/ |
H A D | dummy_hcd.c | 85 struct list_head queue; member in struct:dummy_ep 98 struct list_head queue; /* ep's requests */ member in struct:dummy_request 333 while (!list_empty(&ep->queue)) { 336 req = list_entry(ep->queue.next, struct dummy_request, queue); 337 list_del_init(&req->queue); 670 INIT_LIST_HEAD(&req->queue); 684 WARN_ON(!list_empty(&req->queue)); 702 if (!_req || !list_empty(&req->queue) || !_req->complete) 715 dev_dbg(udc_dev(dum), "ep %p queue re [all...] |
H A D | fsl_qe_udc.c | 82 /* the req->queue pointer is used by ep_queue() func, in which 83 * the request will be added into a udc_ep->queue 'd tail 84 * so here the req will be dropped from the ep->queue 86 list_del_init(&req->queue); 114 /* don't modify queue heads during completion callback */ 131 while (!list_empty(&ep->queue)) { 133 req = list_entry(ep->queue.next, struct qe_req, queue); 780 if (ep->has_data <= 0 && (!list_empty(&ep->queue))) 907 if (list_empty(&ep->queue)) { [all...] |
H A D | mv_u3d_core.c | 180 dev_dbg(u3d->dev, "mv_u3d_done: remove req->queue\n"); 181 /* Removed the req from ep queue */ 182 list_del_init(&req->queue); 235 if (!list_empty(&ep->queue)) { 236 dev_err(u3d->dev, "add trb to non-empty queue!\n"); 505 /* and push them to device queue */ 510 /* irq handler advances the queue */ 511 list_add_tail(&req->queue, &ep->queue); 674 INIT_LIST_HEAD(&req->queue); [all...] |
/linux-master/drivers/scsi/ |
H A D | ips.c | 62 /* - Send all of the commands on the queue at once rather than */ 103 /* - Don't release HA Lock in ips_next() until SC taken off queue */ 798 /* See if the command is on the copp queue */ 808 /* See if the command is on the wait queue */ 1173 /* Set queue depths on devices once scan is complete */ 2503 /* Take the next command off the queue and send it to the controller */ 2522 * Block access to the queue function so 2627 spin_unlock(host->host_lock); /* Unlock HA after command is taken off queue */ 2732 /* Add an item to the head of the queue */ 2738 ips_putq_scb_head(ips_scb_queue_t * queue, ips_scb_ argument 2766 ips_removeq_scb_head(ips_scb_queue_t * queue) argument 2801 ips_removeq_scb(ips_scb_queue_t * queue, ips_scb_t * item) argument 2846 ips_putq_wait_tail(ips_wait_queue_entry_t *queue, struct scsi_cmnd *item) argument 2877 ips_removeq_wait_head(ips_wait_queue_entry_t *queue) argument 2911 ips_removeq_wait(ips_wait_queue_entry_t *queue, struct scsi_cmnd *item) argument 2958 ips_putq_copp_tail(ips_copp_queue_t * queue, ips_copp_wait_item_t * item) argument 2990 ips_removeq_copp_head(ips_copp_queue_t * queue) argument 3025 ips_removeq_copp(ips_copp_queue_t * queue, ips_copp_wait_item_t * item) argument [all...] |
/linux-master/block/ |
H A D | genhd.c | 402 if (queue_is_mq(disk->queue) && disk->fops->poll_bio) 406 * The disk queue should now be all set with enough information about 408 * elevator if one is needed, that is, for devices requesting queue 411 elevator_init_mq(disk->queue); 537 rq_qos_exit(disk->queue); 552 if (disk->queue->elevator) 553 elevator_exit(disk->queue); 593 blk_queue_flag_set(QUEUE_FLAG_DYING, disk->queue); 603 blk_queue_start_drain(disk->queue); 641 struct request_queue *q = disk->queue; [all...] |
H A D | blk-cgroup.c | 162 * Both ->pd_free_fn() and request queue's release handler may 195 * example, don't try to follow throtl_data and request queue links. 311 blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node); 319 if (!blk_get_queue(disk->queue)) 322 blkg->q = disk->queue; 341 if (!blkcg_policy_enabled(disk->queue, pol)) 360 blk_put_queue(disk->queue); 380 lockdep_assert_held(&disk->queue->queue_lock); 383 if (blk_queue_dying(disk->queue)) { 406 blkg->parent = blkg_lookup(blkcg_parent(blkcg), disk->queue); [all...] |
H A D | blk-zoned.c | 207 need_reset = blk_alloc_zone_bitmap(disk->queue->node, disk->nr_zones); 616 zwplug->wp_offset = sector & (disk->queue->limits.chunk_sectors - 1); 640 struct request_queue *q = zwplug->disk->queue; 822 sector += disk->queue->limits.chunk_sectors) { 837 * Grab an extra reference on the BIO request queue usage counter. 842 percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); 956 * Drop the extra reference on the queue usage we got when 1324 * blk-mq devices will reuse the extra reference on the request queue 1330 blk_queue_exit(bdev->bd_disk->queue); 1402 * plugged BIOs to complete if there is a queue freez [all...] |
/linux-master/drivers/usb/gadget/udc/bdc/ |
H A D | bdc_ep.c | 530 list_add_tail(&req->queue, &ep->queue); 547 list_del(&req->queue); 574 /* de-queue any pending requests */ 575 while (!list_empty(&ep->queue)) { 576 req = list_entry(ep->queue.next, struct bdc_req, 577 queue); 828 first_req = list_first_entry(&ep->queue, struct bdc_req, 829 queue); 974 if (unlikely(list_empty(&ep->queue))) { [all...] |
/linux-master/drivers/md/ |
H A D | dm-cache-policy-smq.c | 59 * from the queue itself. 246 * The stochastic-multi-queue is a set of lru lists stacked into levels. 253 struct queue { struct 270 static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels) 288 static unsigned int q_size(struct queue *q) 296 static void q_push(struct queue *q, struct entry *e) 306 static void q_push_front(struct queue *q, struct entry *e) 316 static void q_push_before(struct queue *q, struct entry *old, struct entry *e) 326 static void q_del(struct queue *q, struct entry *e) 336 static struct entry *q_peek(struct queue * [all...] |
/linux-master/drivers/net/wireless/intel/iwlwifi/queue/ |
H A D | tx.c | 14 #include "queue/tx.h" 181 * If skb is not NULL, it means that the whole queue is being 182 * freed and that the queue is not empty - free the skb 701 * should always be less than max_tfd_queue_size elements in the queue. 703 * to reserve any queue entries for this purpose. 733 "queue %d out of range", txq_id)) 737 "TX on unused queue %d\n", txq_id)) 774 /* Set up first empty entry in queue's array of Tx/cmd buffers */ 800 /* start timer if queue currently empty */ 848 /* just in case - this queue ma 1294 iwl_txq_dyn_free(struct iwl_trans *trans, int queue) argument 1308 "queue %d not used", queue); local 1334 struct iwl_txq *queue; local 1720 int queue; local [all...] |
/linux-master/drivers/atm/ |
H A D | nicstar.c | 228 while ((hb = skb_dequeue(&card->hbpool.queue)) != NULL) { 236 while ((iovb = skb_dequeue(&card->iovpool.queue)) != NULL) { 241 while ((lb = skb_dequeue(&card->lbpool.queue)) != NULL) 243 while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL) 655 skb_queue_head_init(&card->hbpool.queue); 669 skb_queue_tail(&card->hbpool.queue, hb); 674 skb_queue_head_init(&card->lbpool.queue); 688 skb_queue_tail(&card->lbpool.queue, lb); 710 skb_queue_head_init(&card->sbpool.queue); 724 skb_queue_tail(&card->sbpool.queue, s [all...] |
/linux-master/ipc/ |
H A D | sem.c | 129 /* One queue for each sleeping process in the system. */ 131 struct list_head list; /* queue of pending operations */ 223 * 2) queue.status: (SEM_BARRIER_2) 305 * merge_queues - merge single semop queues into global queue 308 * This function merges all per-semaphore queues into the global queue. 810 * update_queue is O(N^2) when it restarts scanning the whole queue of 833 * in the queue, then they have observed the original 844 * @wake_q: lockless wake-queue head. 871 /* operation completed, remove from queue & wakeup */ 887 * @wake_q: lockless wake-queue hea 1993 struct sem_queue queue; local [all...] |
/linux-master/drivers/spi/ |
H A D | spi-topcliff-pch.c | 128 * @work: Reference to work queue handler 129 * @wait: Wait queue for waking up upon receiving an 134 * @queue: SPI Message queue 170 struct list_head queue; member in struct:pch_spi_data 482 /* add message to queue */ 483 list_add_tail(&pmsg->queue, &data->queue); 489 dev_dbg(&pspi->dev, "%s - Invoked queue work\n", __func__); 562 /* flush queue an [all...] |
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_ethtool.c | 632 struct ethtool_coalesce *coal, u16 queue) 638 if (queue == COALESCE_ALL_QUEUE) { 643 if (queue >= nic_dev->num_qps) { 645 "Invalid queue_id: %d\n", queue); 648 rx_intr_coal_info = &nic_dev->rx_intr_coalesce[queue]; 649 tx_intr_coal_info = &nic_dev->tx_intr_coalesce[queue]; 710 "Failed to set %s queue%d coalesce", 718 u16 queue, bool set_rx_coal) 723 if (queue == COALESCE_ALL_QUEUE) { 731 if (queue > 631 __hinic_get_coalesce(struct net_device *netdev, struct ethtool_coalesce *coal, u16 queue) argument 716 __set_hw_coal_param(struct hinic_dev *nic_dev, struct hinic_intr_coal_info *intr_coal, u16 queue, bool set_rx_coal) argument 745 __hinic_set_coalesce(struct net_device *netdev, struct ethtool_coalesce *coal, u16 queue) argument 814 hinic_get_per_queue_coalesce(struct net_device *netdev, u32 queue, struct ethtool_coalesce *coal) argument 820 hinic_set_per_queue_coalesce(struct net_device *netdev, u32 queue, struct ethtool_coalesce *coal) argument [all...] |
/linux-master/drivers/scsi/bfa/ |
H A D | bfa_core.c | 700 * Callback only as long as there is room in request queue 766 int queue; local 773 * RME completion queue interrupt 777 for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++) 778 bfa_isr_rspq(bfa, queue); 786 * CPE completion queue interrupt 790 for (queue = 0; queue < BFI_IOC_MAX_CQ 804 int queue; local [all...] |
/linux-master/drivers/media/platform/rockchip/rkisp1/ |
H A D | rkisp1-capture.c | 742 if (!list_empty(&cap->buf.queue)) { 745 cap->buf.next = list_first_entry(&cap->buf.queue, struct rkisp1_buffer, queue); 746 list_del(&cap->buf.next->queue); 867 static int rkisp1_vb2_queue_setup(struct vb2_queue *queue, argument 873 struct rkisp1_capture *cap = queue->drv_priv; 934 list_add_tail(&ispbuf->queue, &cap->buf.queue); 972 while (!list_empty(&cap->buf.queue)) { 973 buf = list_first_entry(&cap->buf.queue, 1123 rkisp1_vb2_stop_streaming(struct vb2_queue *queue) argument 1149 rkisp1_vb2_start_streaming(struct vb2_queue *queue, unsigned int count) argument [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | ap_bus.c | 78 /* Hashtable of all queue devices on the AP bus */ 320 * ap_queue_info(): Check and get AP queue info. 324 * -1 invalid APQN, TAPQ error or AP queue status which 786 int rc, card, queue, devres, drvres; local 790 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); 793 test_bit_inv(queue, ap_perms.aqm); 798 pr_debug("%s reprobing queue=%02x.%04x\n", 799 __func__, card, queue); 802 AP_DBF_WARN("%s reprobing queue=%02x.%04x failed\n", 803 __func__, card, queue); 826 ap_owned_by_def_drv(int card, int queue) argument 856 int card, queue, rc = 0; local 874 int card, queue, devres, drvres, rc = -ENODEV; local [all...] |
/linux-master/drivers/net/wireless/ath/ath5k/ |
H A D | ath5k.h | 138 /* Access QCU registers per queue */ 534 * @AR5K_TX_QUEUE_DATA: A normal data queue 535 * @AR5K_TX_QUEUE_BEACON: The beacon queue 536 * @AR5K_TX_QUEUE_CAB: The after-beacon queue 537 * @AR5K_TX_QUEUE_UAPSD: Unscheduled Automatic Power Save Delivery queue 571 * @AR5K_TX_QUEUE_ID_NOQCU_DATA: Data queue on AR5210 (no QCU available) 572 * @AR5K_TX_QUEUE_ID_NOQCU_BEACON: Beacon queue on AR5210 (no QCU available) 573 * @AR5K_TX_QUEUE_ID_DATA_MIN: Data queue min index 574 * @AR5K_TX_QUEUE_ID_DATA_MAX: Data queue max index 575 * @AR5K_TX_QUEUE_ID_CAB: Content after beacon queue [all...] |
/linux-master/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-dev.c | 463 unsigned int queue) 468 /* Does this queue handle the priority? */ 469 if (pdata->prio2q_map[prio] != queue) 531 /* From MAC ver 30H the TFCR is per priority, instead of per queue */ 2331 unsigned int queue, 2339 if (pdata->pfcq[queue] && (q_fifo_size > pdata->pfc_rfa)) { 2340 /* PFC is active for this queue */ 2355 pdata->rx_rfa[queue] = 0; 2356 pdata->rx_rfd[queue] = 0; 2362 pdata->rx_rfa[queue] 462 xgbe_is_pfc_queue(struct xgbe_prv_data *pdata, unsigned int queue) argument 2330 xgbe_queue_flow_control_threshold(struct xgbe_prv_data *pdata, unsigned int queue, unsigned int q_fifo_size) argument 2659 unsigned int qptc, qptc_extra, queue; local 2740 unsigned int offset, queue, prio; local 3270 xgbe_txq_prepare_tx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument 3296 xgbe_prepare_tx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument 3384 xgbe_prepare_rx_stop(struct xgbe_prv_data *pdata, unsigned int queue) argument [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_priv.h | 138 * eventually HQD registers when the queue is mapped by HWS. In order to 320 * page used by kernel queue 386 * enum kfd_unmap_queues_filter - Enum for queue filters. 405 * enum kfd_queue_type - Enum for various queue types. 407 * @KFD_QUEUE_TYPE_COMPUTE: Regular user mode queue type. 409 * @KFD_QUEUE_TYPE_SDMA: SDMA user mode queue type. 411 * @KFD_QUEUE_TYPE_HIQ: HIQ queue type. 413 * @KFD_QUEUE_TYPE_DIQ: DIQ queue type. 415 * @KFD_QUEUE_TYPE_SDMA_XGMI: Special SDMA queue for XGMI interface. 438 * @type: The queue typ 581 struct queue { struct 590 uint32_t queue; member in struct:queue [all...] |
/linux-master/drivers/s390/net/ |
H A D | qeth_core.h | 186 /* QDIO queue and buffer handling */ 529 static inline void qeth_tx_arm_timer(struct qeth_qdio_out_q *queue, argument 532 timer_reduce(&queue->timer, usecs_to_jiffies(usecs) + jiffies); 535 static inline bool qeth_out_queue_is_full(struct qeth_qdio_out_q *queue) argument 537 return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q; 540 static inline bool qeth_out_queue_is_empty(struct qeth_qdio_out_q *queue) argument 542 return atomic_read(&queue->used_buffers) == 0; 913 struct qeth_qdio_out_q *queue) 915 return qeth_iqd_translate_txq(card->dev, queue->queue_no) == 1103 struct qeth_qdio_out_q *queue, __be1 912 qeth_iqd_is_mcast_queue(struct qeth_card *card, struct qeth_qdio_out_q *queue) argument [all...] |
/linux-master/drivers/staging/media/ipu3/ |
H A D | ipu3.c | 90 * Do not enable dummy buffers for master queue, 115 /* Allocate a dummy buffer for each queue where buffer is optional */ 154 int queue, unsigned int pipe) 160 if (queue == IPU3_CSS_QUEUE_IN) 163 if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr)) 168 if (imgu_css_buf_state(&imgu_pipe->queues[queue].dummybufs[i]) != 175 imgu_css_buf_init(&imgu_pipe->queues[queue].dummybufs[i], queue, 176 imgu_pipe->queues[queue].dmap.daddr); 178 return &imgu_pipe->queues[queue] 153 imgu_dummybufs_get(struct imgu_device *imgu, int queue, unsigned int pipe) argument [all...] |
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_req.c | 45 struct rxe_queue *q = qp->sq.queue; 58 wqe = queue_addr_from_index(qp->sq.queue, wqe_index); 108 /* request a send queue retry */ 126 q = qp->sq.queue; 164 struct rxe_queue *q = qp->sq.queue; 200 * @qp: the queue pair 209 * Requires ALL previous operations on the send queue 213 return qp->req.wqe_index != queue_get_consumer(qp->sq.queue, 583 qp->req.wqe_index = queue_next_index(qp->sq.queue, 634 qp->req.wqe_index = queue_next_index(qp->sq.queue, q [all...] |
/linux-master/fs/ocfs2/dlm/ |
H A D | dlmrecovery.c | 214 * 5) the new master collects up all of secondary lock queue info 217 * 6) each secondary lock queue responds with the full known lock info 220 * 8) upon receiving this message, the secondary queue node unlocks 849 /* queue up work for dlm_request_all_locks_worker */ 1083 struct list_head *iter, *queue = &res->granted; local 1087 list_for_each(iter, queue) 1089 queue++; 1173 int queue) 1179 if (queue == DLM_BLOCKED_LIST) 1208 struct dlm_migratable_lockres *mres, int queue) 1171 dlm_prepare_lvb_for_migration(struct dlm_lock *lock, struct dlm_migratable_lockres *mres, int queue) argument 1207 dlm_add_lock_to_array(struct dlm_lock *lock, struct dlm_migratable_lockres *mres, int queue) argument 1264 struct list_head *queue; local 1792 struct list_head *queue, *iter; local 2048 struct list_head *queue; local 2212 struct list_head *queue; local [all...] |
/linux-master/drivers/block/ |
H A D | virtio_blk.c | 134 struct virtio_blk *vblk = hctx->queue->queuedata; 230 return blk_rq_map_sg(hctx->queue, req, vbr->sg_table.sgl); 339 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; 372 /* In case queue is stopped waiting for more buffers. */ 374 blk_mq_start_stopped_hw_queues(vblk->disk->queue, true); 380 struct virtio_blk *vblk = hctx->queue->queuedata; 430 struct virtio_blk *vblk = hctx->queue->queuedata; 447 /* Don't stop the queue if -ENOMEM: we may have failed to 468 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; 537 struct request_queue *q = vblk->disk->queue; [all...] |