Searched refs:queue (Results 551 - 575 of 1381) sorted by relevance

<<21222324252627282930>>

/linux-master/drivers/input/touchscreen/
H A Dsur40.c217 struct vb2_queue queue; member in struct:sur40_state
537 if (!vb2_start_streaming_called(&sur40->queue))
543 dev_dbg(sur40->dev, "buffer queue empty\n");
746 sur40->queue = sur40_queue;
747 sur40->queue.drv_priv = sur40;
748 sur40->queue.lock = &sur40->lock;
749 sur40->queue.dev = sur40->dev;
751 /* initialize the queue */
752 error = vb2_queue_init(&sur40->queue);
760 sur40->vdev.queue
[all...]
/linux-master/drivers/net/ethernet/microsoft/mana/
H A Dhw_channel.c259 struct gdma_queue **queue)
270 return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
277 struct gdma_queue **queue)
288 return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
294 struct gdma_queue **queue)
306 return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
488 struct gdma_queue *queue; local
507 err = mana_hwc_create_gdma_wq(hwc, q_type, queue_size, &queue);
512 hwc_wq->gdma_wq = queue;
668 /* CQ is shared by SQ and RQ, so CQ's queue dept
257 mana_hwc_create_gdma_wq(struct hw_channel_context *hwc, enum gdma_queue_type type, u64 queue_size, struct gdma_queue **queue) argument
273 mana_hwc_create_gdma_cq(struct hw_channel_context *hwc, u64 queue_size, void *ctx, gdma_cq_callback *cb, struct gdma_queue *parent_eq, struct gdma_queue **queue) argument
291 mana_hwc_create_gdma_eq(struct hw_channel_context *hwc, u64 queue_size, void *ctx, gdma_eq_callback *cb, struct gdma_queue **queue) argument
[all...]
/linux-master/include/linux/
H A Diommu.h136 * struct iopf_queue - IO Page Fault queue
138 * @devices: devices attached to this queue
165 #define __IOMMU_DOMAIN_DMA_FQ (1U << 3) /* DMA-API uses flush queue */
248 * usefully support the non-strict DMA flush queue.
617 * queue
688 * @queue: IOPF queue
689 * @queue_list: index into queue->devices
700 struct iopf_queue *queue; member in struct:iommu_fault_param
1565 int iopf_queue_add_device(struct iopf_queue *queue, struc
1577 iopf_queue_add_device(struct iopf_queue *queue, struct device *dev) argument
1583 iopf_queue_remove_device(struct iopf_queue *queue, struct device *dev) argument
1597 iopf_queue_free(struct iopf_queue *queue) argument
1601 iopf_queue_discard_partial(struct iopf_queue *queue) argument
[all...]
/linux-master/drivers/usb/chipidea/
H A Dudc.c556 /* don't queue twice */
590 if (!list_empty(&hwep->qh.queue)) {
597 hwreqprev = list_entry(hwep->qh.queue.prev,
598 struct ci_hw_req, queue);
761 while (!list_empty(&hwep->qh.queue)) {
764 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next,
765 struct ci_hw_req, queue);
774 list_del_init(&hwreq->queue);
805 !list_empty(&hwep->qh.queue) &&
944 if (!list_empty(&hwep->qh.queue)) {
[all...]
/linux-master/drivers/scsi/
H A Dipr.c679 * @hrrq: hrr queue
691 struct ipr_cmnd, queue);
692 list_del(&ipr_cmd->queue);
822 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
863 temp, &hrrq->hrrq_pending_q, queue) {
864 list_del(&ipr_cmd->queue);
930 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
1051 * place the hostrcb on the free queue.
1064 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
1065 list_add_tail(&hostrcb->queue,
[all...]
/linux-master/tools/perf/util/
H A Dintel-pt.c456 struct auxtrace_queue *queue; local
459 queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
464 buffer = auxtrace_buffer__next(queue, buffer);
501 struct auxtrace_queue *queue; local
509 queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
511 buffer = auxtrace_buffer__next(queue, buffer);
1421 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; local
1422 struct intel_pt_queue *ptq = queue->priv;
1479 struct auxtrace_queue *queue)
1481 struct intel_pt_queue *ptq = queue
1478 intel_pt_set_pid_tid_cpu(struct intel_pt *pt, struct auxtrace_queue *queue) argument
1573 intel_pt_setup_queue(struct intel_pt *pt, struct auxtrace_queue *queue, unsigned int queue_nr) argument
2546 struct auxtrace_queue *queue; local
2817 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; local
2832 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; local
3014 struct auxtrace_queue *queue; local
3070 struct auxtrace_queue *queue = &pt->queues.queue_array[i]; local
3082 intel_pt_sample_set_pid_tid_cpu(struct intel_pt_queue *ptq, struct auxtrace_queue *queue, struct perf_sample *sample) argument
3113 struct auxtrace_queue *queue; local
3363 struct auxtrace_queue *queue; local
[all...]
H A Dannotate.c698 int max_lines, struct annotation_line *queue, int addr_fmt_width,
729 if (queue != NULL) {
730 list_for_each_entry_from(queue, &notes->src->source, node) {
731 if (queue == al)
733 annotation_line__print(queue, sym, start, evsel, len,
780 if (queue)
1102 struct annotation_line *pos, *queue = NULL; local
1148 if (context && queue == NULL) {
1149 queue = pos;
1155 queue, addr_fmt_widt
696 annotation_line__print(struct annotation_line *al, struct symbol *sym, u64 start, struct evsel *evsel, u64 len, int min_pcnt, int printed, int max_lines, struct annotation_line *queue, int addr_fmt_width, int percent_type) argument
2462 struct list_head queue; member in struct:basic_block_data
[all...]
/linux-master/drivers/net/wireless/ath/ath6kl/
H A Dhtc_mbox.c600 struct list_head *queue)
616 "htc tx got packet 0x%p queue depth %d\n",
626 /* now we can fully move onto caller's queue */
629 list_move_tail(&packet->list, queue);
682 struct list_head *queue)
694 if (list_empty(queue))
697 packet = list_first_entry(queue, struct htc_packet, list);
709 /* now remove it from the queue */
736 list_add(&packet->list, queue);
746 * Drain a queue an
598 ath6kl_htc_tx_pkts_get(struct htc_target *target, struct htc_endpoint *endpoint, struct list_head *queue) argument
678 ath6kl_htc_tx_setup_scat_list(struct htc_target *target, struct htc_endpoint *endpoint, struct hif_scatter_req *scat_req, int n_scat, struct list_head *queue) argument
754 ath6kl_htc_tx_bundle(struct htc_endpoint *endpoint, struct list_head *queue, int *sent_bundle, int *n_bundle_pkts) argument
1124 struct list_head queue; local
1270 struct list_head queue; local
1365 ath6kl_htc_rx_setup(struct htc_target *target, struct htc_endpoint *ep, u32 *lk_ahds, struct list_head *queue, int n_msg) argument
1468 ath6kl_htc_rx_alloc(struct htc_target *target, u32 lk_ahds[], int msg, struct htc_endpoint *endpoint, struct list_head *queue) argument
[all...]
/linux-master/drivers/usb/gadget/udc/
H A Dsnps_udc_core.c454 INIT_LIST_HEAD(&ep->queue);
531 INIT_LIST_HEAD(&req->queue);
592 BUG_ON(!list_empty(&req->queue));
981 /* remove from ep queue */
982 list_del_init(&req->queue);
1062 || !list_empty(&req->queue))
1083 VDBG(dev, "%s queue req %p, len %d req->td_data=%p buf %p\n",
1092 /* on empty queue just do first transfer */
1093 if (list_empty(&ep->queue)) {
1197 /* add request to ep queue */
[all...]
H A Drenesas_usbf.c260 struct list_head queue; member in struct:usbf_req
278 struct list_head queue; member in struct:usbf_ep
1421 /* Process queue at bridge interrupt only */
1540 list_del_init(&req->queue);
1569 while (!list_empty(&ep->queue)) {
1570 req = list_first_entry(&ep->queue, struct usbf_req, queue);
1604 req = list_first_entry_or_null(&epn->queue, struct usbf_req, queue);
1664 req = list_first_entry_or_null(&ep->queue, struc
[all...]
H A Dr8a66597-udc.h52 struct list_head queue; member in struct:r8a66597_request
60 struct list_head queue; member in struct:r8a66597_ep
H A Dmv_udc.h228 struct list_head queue; member in struct:mv_ep
243 struct list_head queue; member in struct:mv_req
H A Dgoku_udc.h215 struct list_head queue; member in struct:goku_ep
224 struct list_head queue; member in struct:goku_request
H A Dmv_u3d.h290 struct list_head queue; /* ep request queued hardware */ member in struct:mv_u3d_ep
308 struct list_head queue; /* ep requst queued on hardware */ member in struct:mv_u3d_req
H A Datmel_usba_udc.h221 #define DBG_QUEUE 0x0020 /* debug request queue processing */
277 struct list_head queue; member in struct:usba_ep
301 struct list_head queue; member in struct:usba_request
/linux-master/drivers/staging/rtl8723bs/core/
H A Drtw_cmd.c167 INIT_LIST_HEAD(&pcmdpriv->cmd_queue.queue);
251 int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj) argument
258 /* spin_lock_bh(&queue->lock); */
259 spin_lock_irqsave(&queue->lock, irqL);
261 list_add_tail(&obj->list, &queue->queue);
263 /* spin_unlock_bh(&queue->lock); */
264 spin_unlock_irqrestore(&queue->lock, irqL);
270 struct cmd_obj *_rtw_dequeue_cmd(struct __queue *queue) argument
275 /* spin_lock_bh(&(queue
[all...]
/linux-master/drivers/net/ethernet/ti/
H A Dam65-cpsw-qos.c1201 int queue, u32 rate_mbps)
1211 queue, rate_mbps, common->tx_ch_rate_msk);
1218 ch_rate = netdev_get_tx_queue(ndev, queue)->tx_maxrate;
1230 if (rate_mbps && !(tx_ch_rate_msk_new & BIT(queue))) {
1231 tx_ch_rate_msk_new |= BIT(queue);
1232 ch_msk = GENMASK(common->tx_ch_num - 1, queue);
1235 tx_ch_rate_msk_new &= ~BIT(queue);
1236 ch_msk = queue ? GENMASK(queue - 1, 0) : 0;
1247 tx_chn = &common->tx_chns[queue];
1200 am65_cpsw_qos_ndo_tx_p0_set_maxrate(struct net_device *ndev, int queue, u32 rate_mbps) argument
[all...]
/linux-master/drivers/infiniband/sw/rxe/
H A Drxe_verbs.c478 attr->max_wr = srq->rq.queue->buf->index_mask;
871 full = queue_full(sq->queue, QUEUE_TYPE_FROM_ULP);
873 rxe_err_qp(qp, "send queue full\n");
877 send_wqe = queue_producer_addr(sq->queue, QUEUE_TYPE_FROM_ULP);
880 queue_advance_producer(sq->queue, QUEUE_TYPE_FROM_ULP);
958 full = queue_full(rq->queue, QUEUE_TYPE_FROM_ULP);
961 rxe_dbg("queue full\n");
982 recv_wqe = queue_producer_addr(rq->queue, QUEUE_TYPE_FROM_ULP);
993 queue_advance_producer(rq->queue, QUEUE_TYPE_FROM_ULP);
1153 cqe = queue_head(cq->queue, QUEUE_TYPE_TO_UL
[all...]
/linux-master/tools/tracing/latency/
H A Dlatency-collector.c127 "Could not queue trace for printing. It is likely that events happen faster\n"
221 struct queue { struct
231 struct queue printqueue;
787 static void init_queue(struct queue *q)
797 static __always_inline int queue_len(const struct queue *q)
805 static __always_inline int queue_nr_free(const struct queue *q)
811 * queue is full because adding an element will make it appear empty
823 static __always_inline void queue_push_to_back(struct queue *q,
830 static __always_inline struct entry queue_pop_from_front(struct queue *q)
838 static __always_inline void queue_cond_signal(struct queue *
[all...]
/linux-master/drivers/scsi/cxlflash/
H A Dcommon.h63 /* Hardware queue definitions */
161 struct list_head queue; member in struct:afu_cmd
184 INIT_LIST_HEAD(&afuc->queue);
214 spinlock_t hsq_slock; /* Hardware send queue lock */
/linux-master/drivers/scsi/aacraid/
H A Dsa.c55 aac_command_normal(&dev->queues->queue[HostNormCmdQueue]);
58 aac_response_normal(&dev->queues->queue[HostNormRespQueue]);
/linux-master/net/rose/
H A Drose_link.c167 while ((skbn = skb_dequeue(&neigh->queue)) != NULL)
281 skb_queue_tail(&neigh->queue, skb);
/linux-master/drivers/i2c/busses/
H A Di2c-octeon-platdrv.c108 wake_up(&i2c->queue);
190 init_waitqueue_head(&i2c->queue);
/linux-master/block/
H A Dblk-rq-qos.c303 struct request_queue *q = disk->queue;
312 * No IO can be in-flight when adding rqos, so freeze queue, which
313 * is fine since we only support rq_qos for blk-mq queue.
338 struct request_queue *q = rqos->disk->queue;
/linux-master/drivers/net/ethernet/intel/i40e/
H A Di40e_lan_hmc.c101 /* Tx queue context information */
121 /* Rx queue context information */
1038 * i40e_clear_lan_tx_queue_context - clear the HMC context for the queue
1040 * @queue: the queue we care about
1043 u16 queue)
1049 I40E_HMC_LAN_TX, queue);
1057 * i40e_set_lan_tx_queue_context - set the HMC context for the queue
1059 * @queue: the queue w
1042 i40e_clear_lan_tx_queue_context(struct i40e_hw *hw, u16 queue) argument
1062 i40e_set_lan_tx_queue_context(struct i40e_hw *hw, u16 queue, struct i40e_hmc_obj_txq *s) argument
1083 i40e_clear_lan_rx_queue_context(struct i40e_hw *hw, u16 queue) argument
1103 i40e_set_lan_rx_queue_context(struct i40e_hw *hw, u16 queue, struct i40e_hmc_obj_rxq *s) argument
[all...]

Completed in 278 milliseconds

<<21222324252627282930>>