Searched refs:queue (Results 326 - 350 of 1381) sorted by relevance

<<11121314151617181920>>

/linux-master/drivers/gpu/drm/i915/
H A Di915_scheduler_types.h49 * DAG of each request, we are able to insert it into a sorted queue when it
57 * is put to the back of its priority queue, then reshuffling its dependents).
97 * A schedule engine represents a submission queue with different priority
98 * bands. It contains all the common state (relative to the backend) to queue,
140 * When we add requests into the queue, or adjust the priority of
143 * we need to preempt the executing requests to service the queue.
152 * @queue: queue of requests, in priority lists
154 struct rb_root_cached queue; member in struct:i915_sched_engine
/linux-master/drivers/media/platform/xilinx/
H A Dxilinx-dma.h66 * @lock: protects the @format, @fmtinfo and @queue fields
69 * @queue: vb2 buffers queue
91 struct vb2_queue queue; member in struct:xvip_dma
/linux-master/block/
H A Dblk-ia-ranges.c111 struct request_queue *q = disk->queue;
154 struct request_queue *q = disk->queue;
269 GFP_KERNEL, disk->queue->node);
281 * Set the independent access ranges information of the request queue
290 struct request_queue *q = disk->queue;
304 * This may be called for a registered queue. E.g. during a device
307 * queue is not registered, registration of the device request queue
H A Dblk-mq-sysfs.c59 q = hctx->queue;
160 struct request_queue *q = hctx->queue;
221 struct request_queue *q = disk->queue;
258 struct request_queue *q = disk->queue;
H A Dbfq-iosched.h50 * bfq_service_tree. All the fields are protected by the queue lock
73 * bfq_sched_data is the basic scheduler queue. It supports three
74 * ioprio_classes, and can be used either as a toplevel queue or as an
75 * intermediate queue in a hierarchical setup.
81 * queue requests are served according to B-WF2Q+.
87 * queue becomes idle when it expires, and 3) if the entity pointed by
88 * in_service_entity is not a queue, then the in-service child entity
148 * guaranteed. All the fields are protected by the queue lock of the
185 /* weight of the queue */
198 * scheduler queue,
514 struct request_queue *queue; member in struct:bfq_data
[all...]
/linux-master/drivers/usb/gadget/udc/aspeed-vhub/
H A Depn.c93 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
103 * Request not active, move on to processing queue, active request
142 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req,
143 queue);
223 * descriptors request an interrupt before queue empty
264 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
315 req = list_first_entry_or_null(&ep->queue,
317 queue);
[all...]
/linux-master/drivers/block/aoe/
H A Daoechr.c127 struct sk_buff_head queue; local
128 __skb_queue_head_init(&queue);
129 __skb_queue_tail(&queue, skb);
130 aoenet_xmit(&queue);
/linux-master/drivers/s390/cio/
H A Dqdio.h88 static inline int do_sqbs(u64 token, unsigned char state, int queue, argument
91 unsigned long _queuestart = ((unsigned long)queue << 32) | *start;
106 static inline int do_eqbs(u64 token, unsigned char *state, int queue, argument
109 unsigned long _queuestart = ((unsigned long)queue << 32) | *start;
164 /* Batch of SBALs that we processed while polling the queue: */
193 /* last scan of the queue */
200 /* queue number */
203 /* bitmask of queue number */
206 /* input or output queue */
288 /* the highest iqdio queue i
[all...]
/linux-master/include/net/
H A Dfq.h17 * struct fq_flow - per traffic flow queue
26 * @queue: sk_buff queue to hold packets
27 * @backlog: number of bytes pending in the queue. The number of packets can be
28 * found in @queue.qlen
34 struct sk_buff_head queue; member in struct:fq_flow
H A Dfq_impl.h54 skb = __skb_dequeue(&flow->queue);
73 pending = min_t(int, 32, skb_queue_len(&flow->queue) / 2);
75 skb = __skb_dequeue(&flow->queue);
225 __skb_queue_tail(&flow->queue, skb);
263 skb_queue_walk_safe(&flow->queue, skb, tmp) {
267 __skb_unlink(skb, &flow->queue);
338 __skb_queue_head_init(&flow->queue);
/linux-master/drivers/crypto/qce/
H A Dcore.c93 ret = crypto_enqueue_request(&qce->queue, req);
101 backlog = crypto_get_backlog(&qce->queue);
102 async_req = crypto_dequeue_request(&qce->queue);
255 crypto_init_queue(&qce->queue, QCE_QUEUE_LENGTH);
/linux-master/drivers/media/pci/mgb4/
H A Dmgb4_vin.h42 struct vb2_queue queue; member in struct:mgb4_vin_dev
45 spinlock_t qlock; /* video buffer queue lock */
H A Dmgb4_vout.h39 struct vb2_queue queue; member in struct:mgb4_vout_dev
42 spinlock_t qlock; /* buffer queue lock */
/linux-master/drivers/staging/media/starfive/camss/
H A Dstf-video.h46 struct list_head queue; member in struct:stfcamss_buffer
79 struct mutex q_lock; /* protects the queue */
/linux-master/arch/mips/include/asm/octeon/
H A Dcvmx-cmd-queue.h33 * The common command queue infrastructure abstracts out the
34 * software necessary for adding to Octeon's chained queue
40 * call cvmx-cmd-queue functions directly. Instead the hardware
46 * cvmx-cmd-queue, knowledge of its internal working can help
49 * Command queue pointers are stored in a global named block
51 * hardware queue is stored in its own cache line to reduce SMP
53 * every 16th queue is next to each other in memory. This scheme
56 * the first queue for each port is in the same cache area. The
63 * In addition to the memory pointer layout, cvmx-cmd-queue
84 * don't use it and it slows down the command queue processin
[all...]
/linux-master/drivers/md/dm-vdo/
H A Dcompletion.c25 * only be enqueued on one queue or operating in a single zone at a time.
123 vdo_enqueue_work_queue(vdo->threads[thread_id].queue, completion);
/linux-master/include/crypto/internal/
H A Dhash.h222 static inline int ahash_enqueue_request(struct crypto_queue *queue, argument
225 return crypto_enqueue_request(queue, &request->base);
229 struct crypto_queue *queue)
231 return ahash_request_cast(crypto_dequeue_request(queue));
228 ahash_dequeue_request( struct crypto_queue *queue) argument
/linux-master/drivers/md/
H A Ddm-zone.c134 struct request_queue *q = md->queue;
184 disk->queue->limits.max_open_zones = 0;
185 disk->queue->limits.max_active_zones = 0;
186 disk->queue->limits.zoned = false;
261 * mapped device queue as needing zone append emulation.
/linux-master/drivers/gpu/drm/panfrost/
H A Dpanfrost_job.c37 struct panfrost_queue_state queue[NUM_JOB_SLOTS]; member in struct:panfrost_job_slot
53 int queue; member in struct:panfrost_fence
71 switch (f->queue) {
98 fence->queue = js_num;
99 fence->seqno = ++js->queue[js_num].emit_seqno;
101 js->queue[js_num].fence_context, fence->seqno);
244 /* Don't queue the job if a reset is in progress */
476 drm_sched_fault(&pfdev->js->queue[js].sched);
656 drm_sched_stop(&pfdev->js->queue[i].sched, bad);
725 drm_sched_resubmit_jobs(&pfdev->js->queue[
[all...]
/linux-master/drivers/usb/gadget/function/
H A Duvc_v4l2.c245 if (fmt->type != video->queue.queue.type)
404 if (b->type != video->queue.queue.type)
407 return uvcg_alloc_buffers(&video->queue, b);
417 return uvcg_query_buffer(&video->queue, b);
428 ret = uvcg_queue_buffer(&video->queue, b);
445 return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK);
456 if (type != video->queue.queue
[all...]
/linux-master/drivers/char/ipmi/
H A Dbt-bmc.c62 wait_queue_head_t queue; member in struct:bt_bmc
187 if (wait_event_interruptible(bt_bmc->queue,
261 if (wait_event_interruptible(bt_bmc->queue,
325 poll_wait(file, &bt_bmc->queue, wait);
353 wake_up(&bt_bmc->queue);
371 wake_up(&bt_bmc->queue);
427 init_waitqueue_head(&bt_bmc->queue);
/linux-master/drivers/net/ethernet/chelsio/cxgb4/
H A Dcxgb4_uld.h255 unsigned int queue);
258 unsigned int queue);
260 unsigned int queue, bool ipv6);
263 unsigned int queue,
266 unsigned int queue, bool ipv6);
299 static inline void set_wr_txq(struct sk_buff *skb, int prio, int queue) argument
301 skb_set_queue_mapping(skb, (queue << 1) | prio);
429 const unsigned short *rxq_ids; /* the ULD's Rx queue ids */
448 /* scheduler queue */
/linux-master/drivers/staging/rtl8723bs/core/
H A Drtw_recv.c28 INIT_LIST_HEAD(&psta_recvpriv->defrag_q.queue);
40 INIT_LIST_HEAD(&precvpriv->free_recv_queue.queue);
42 INIT_LIST_HEAD(&precvpriv->recv_pending_queue.queue);
44 INIT_LIST_HEAD(&precvpriv->uc_swdec_pending_queue.queue);
68 list_add_tail(&(precvframe->u.list), &(precvpriv->free_recv_queue.queue));
113 if (list_empty(&pfree_recv_queue->queue))
173 signed int _rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue) argument
183 list_add_tail(&(precvframe->u.hdr.list), get_list_head(queue));
186 if (queue == &precvpriv->free_recv_queue)
192 signed int rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue) argument
248 rtw_enqueue_recvbuf_to_head(struct recv_buf *precvbuf, struct __queue *queue) argument
260 rtw_enqueue_recvbuf(struct recv_buf *precvbuf, struct __queue *queue) argument
272 rtw_dequeue_recvbuf(struct __queue *queue) argument
[all...]
/linux-master/drivers/infiniband/sw/rdmavt/
H A Dcq.c15 * rvt_cq_enter - add a new entry to the completion queue
16 * @cq: completion queue
39 u_wc = cq->queue;
149 * rvt_create_cq - create a completion queue
183 * Allocate the completion queue entries and head/tail pointers.
254 cq->queue = u_wc;
270 * rvt_destroy_cq - destroy a completion queue
271 * @ibcq: the completion queue to destroy.
293 * rvt_req_notify_cq - change the notification type for a completion queue
294 * @ibcq: the completion queue
[all...]
/linux-master/drivers/accel/habanalabs/common/
H A Dirq.c29 * @ptr: the current ci or pi value of the completion queue
31 * Increment ptr by 1. If it reaches the number of completion queue
45 * @ptr: the current ci value of the event queue
47 * Increment ptr by 1. If it reaches the number of event queue
70 * job_finish - queue job finish work
74 * @cq: completion queue
80 struct hl_hw_queue *queue; local
83 queue = &hdev->kernel_queues[cq->hw_queue_id];
84 job = queue->shadow_queue[hl_pi_2_offset(cs_seq)];
88 atomic_inc(&queue
102 struct hl_hw_queue *queue; local
[all...]

Completed in 396 milliseconds

<<11121314151617181920>>