/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_scheduler_types.h | 49 * DAG of each request, we are able to insert it into a sorted queue when it 57 * is put to the back of its priority queue, then reshuffling its dependents). 97 * A schedule engine represents a submission queue with different priority 98 * bands. It contains all the common state (relative to the backend) to queue, 140 * When we add requests into the queue, or adjust the priority of 143 * we need to preempt the executing requests to service the queue. 152 * @queue: queue of requests, in priority lists 154 struct rb_root_cached queue; member in struct:i915_sched_engine
|
/linux-master/drivers/media/platform/xilinx/ |
H A D | xilinx-dma.h | 66 * @lock: protects the @format, @fmtinfo and @queue fields 69 * @queue: vb2 buffers queue 91 struct vb2_queue queue; member in struct:xvip_dma
|
/linux-master/block/ |
H A D | blk-ia-ranges.c | 111 struct request_queue *q = disk->queue; 154 struct request_queue *q = disk->queue; 269 GFP_KERNEL, disk->queue->node); 281 * Set the independent access ranges information of the request queue 290 struct request_queue *q = disk->queue; 304 * This may be called for a registered queue. E.g. during a device 307 * queue is not registered, registration of the device request queue
|
H A D | blk-mq-sysfs.c | 59 q = hctx->queue; 160 struct request_queue *q = hctx->queue; 221 struct request_queue *q = disk->queue; 258 struct request_queue *q = disk->queue;
|
H A D | bfq-iosched.h | 50 * bfq_service_tree. All the fields are protected by the queue lock 73 * bfq_sched_data is the basic scheduler queue. It supports three 74 * ioprio_classes, and can be used either as a toplevel queue or as an 75 * intermediate queue in a hierarchical setup. 81 * queue requests are served according to B-WF2Q+. 87 * queue becomes idle when it expires, and 3) if the entity pointed by 88 * in_service_entity is not a queue, then the in-service child entity 148 * guaranteed. All the fields are protected by the queue lock of the 185 /* weight of the queue */ 198 * scheduler queue, 514 struct request_queue *queue; member in struct:bfq_data [all...] |
/linux-master/drivers/usb/gadget/udc/aspeed-vhub/ |
H A D | epn.c | 93 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue); 103 * Request not active, move on to processing queue, active request 142 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, 143 queue); 223 * descriptors request an interrupt before queue empty 264 req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue); 315 req = list_first_entry_or_null(&ep->queue, 317 queue); [all...] |
/linux-master/drivers/block/aoe/ |
H A D | aoechr.c | 127 struct sk_buff_head queue; local 128 __skb_queue_head_init(&queue); 129 __skb_queue_tail(&queue, skb); 130 aoenet_xmit(&queue);
|
/linux-master/drivers/s390/cio/ |
H A D | qdio.h | 88 static inline int do_sqbs(u64 token, unsigned char state, int queue, argument 91 unsigned long _queuestart = ((unsigned long)queue << 32) | *start; 106 static inline int do_eqbs(u64 token, unsigned char *state, int queue, argument 109 unsigned long _queuestart = ((unsigned long)queue << 32) | *start; 164 /* Batch of SBALs that we processed while polling the queue: */ 193 /* last scan of the queue */ 200 /* queue number */ 203 /* bitmask of queue number */ 206 /* input or output queue */ 288 /* the highest iqdio queue i [all...] |
/linux-master/include/net/ |
H A D | fq.h | 17 * struct fq_flow - per traffic flow queue 26 * @queue: sk_buff queue to hold packets 27 * @backlog: number of bytes pending in the queue. The number of packets can be 28 * found in @queue.qlen 34 struct sk_buff_head queue; member in struct:fq_flow
|
H A D | fq_impl.h | 54 skb = __skb_dequeue(&flow->queue); 73 pending = min_t(int, 32, skb_queue_len(&flow->queue) / 2); 75 skb = __skb_dequeue(&flow->queue); 225 __skb_queue_tail(&flow->queue, skb); 263 skb_queue_walk_safe(&flow->queue, skb, tmp) { 267 __skb_unlink(skb, &flow->queue); 338 __skb_queue_head_init(&flow->queue);
|
/linux-master/drivers/crypto/qce/ |
H A D | core.c | 93 ret = crypto_enqueue_request(&qce->queue, req); 101 backlog = crypto_get_backlog(&qce->queue); 102 async_req = crypto_dequeue_request(&qce->queue); 255 crypto_init_queue(&qce->queue, QCE_QUEUE_LENGTH);
|
/linux-master/drivers/media/pci/mgb4/ |
H A D | mgb4_vin.h | 42 struct vb2_queue queue; member in struct:mgb4_vin_dev 45 spinlock_t qlock; /* video buffer queue lock */
|
H A D | mgb4_vout.h | 39 struct vb2_queue queue; member in struct:mgb4_vout_dev 42 spinlock_t qlock; /* buffer queue lock */
|
/linux-master/drivers/staging/media/starfive/camss/ |
H A D | stf-video.h | 46 struct list_head queue; member in struct:stfcamss_buffer 79 struct mutex q_lock; /* protects the queue */
|
/linux-master/arch/mips/include/asm/octeon/ |
H A D | cvmx-cmd-queue.h | 33 * The common command queue infrastructure abstracts out the 34 * software necessary for adding to Octeon's chained queue 40 * call cvmx-cmd-queue functions directly. Instead the hardware 46 * cvmx-cmd-queue, knowledge of its internal working can help 49 * Command queue pointers are stored in a global named block 51 * hardware queue is stored in its own cache line to reduce SMP 53 * every 16th queue is next to each other in memory. This scheme 56 * the first queue for each port is in the same cache area. The 63 * In addition to the memory pointer layout, cvmx-cmd-queue 84 * don't use it and it slows down the command queue processin [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | completion.c | 25 * only be enqueued on one queue or operating in a single zone at a time. 123 vdo_enqueue_work_queue(vdo->threads[thread_id].queue, completion);
|
/linux-master/include/crypto/internal/ |
H A D | hash.h | 222 static inline int ahash_enqueue_request(struct crypto_queue *queue, argument 225 return crypto_enqueue_request(queue, &request->base); 229 struct crypto_queue *queue) 231 return ahash_request_cast(crypto_dequeue_request(queue)); 228 ahash_dequeue_request( struct crypto_queue *queue) argument
|
/linux-master/drivers/md/ |
H A D | dm-zone.c | 134 struct request_queue *q = md->queue; 184 disk->queue->limits.max_open_zones = 0; 185 disk->queue->limits.max_active_zones = 0; 186 disk->queue->limits.zoned = false; 261 * mapped device queue as needing zone append emulation.
|
/linux-master/drivers/gpu/drm/panfrost/ |
H A D | panfrost_job.c | 37 struct panfrost_queue_state queue[NUM_JOB_SLOTS]; member in struct:panfrost_job_slot 53 int queue; member in struct:panfrost_fence 71 switch (f->queue) { 98 fence->queue = js_num; 99 fence->seqno = ++js->queue[js_num].emit_seqno; 101 js->queue[js_num].fence_context, fence->seqno); 244 /* Don't queue the job if a reset is in progress */ 476 drm_sched_fault(&pfdev->js->queue[js].sched); 656 drm_sched_stop(&pfdev->js->queue[i].sched, bad); 725 drm_sched_resubmit_jobs(&pfdev->js->queue[ [all...] |
/linux-master/drivers/usb/gadget/function/ |
H A D | uvc_v4l2.c | 245 if (fmt->type != video->queue.queue.type) 404 if (b->type != video->queue.queue.type) 407 return uvcg_alloc_buffers(&video->queue, b); 417 return uvcg_query_buffer(&video->queue, b); 428 ret = uvcg_queue_buffer(&video->queue, b); 445 return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK); 456 if (type != video->queue.queue [all...] |
/linux-master/drivers/char/ipmi/ |
H A D | bt-bmc.c | 62 wait_queue_head_t queue; member in struct:bt_bmc 187 if (wait_event_interruptible(bt_bmc->queue, 261 if (wait_event_interruptible(bt_bmc->queue, 325 poll_wait(file, &bt_bmc->queue, wait); 353 wake_up(&bt_bmc->queue); 371 wake_up(&bt_bmc->queue); 427 init_waitqueue_head(&bt_bmc->queue);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_uld.h | 255 unsigned int queue); 258 unsigned int queue); 260 unsigned int queue, bool ipv6); 263 unsigned int queue, 266 unsigned int queue, bool ipv6); 299 static inline void set_wr_txq(struct sk_buff *skb, int prio, int queue) argument 301 skb_set_queue_mapping(skb, (queue << 1) | prio); 429 const unsigned short *rxq_ids; /* the ULD's Rx queue ids */ 448 /* scheduler queue */
|
/linux-master/drivers/staging/rtl8723bs/core/ |
H A D | rtw_recv.c | 28 INIT_LIST_HEAD(&psta_recvpriv->defrag_q.queue); 40 INIT_LIST_HEAD(&precvpriv->free_recv_queue.queue); 42 INIT_LIST_HEAD(&precvpriv->recv_pending_queue.queue); 44 INIT_LIST_HEAD(&precvpriv->uc_swdec_pending_queue.queue); 68 list_add_tail(&(precvframe->u.list), &(precvpriv->free_recv_queue.queue)); 113 if (list_empty(&pfree_recv_queue->queue)) 173 signed int _rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue) argument 183 list_add_tail(&(precvframe->u.hdr.list), get_list_head(queue)); 186 if (queue == &precvpriv->free_recv_queue) 192 signed int rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue) argument 248 rtw_enqueue_recvbuf_to_head(struct recv_buf *precvbuf, struct __queue *queue) argument 260 rtw_enqueue_recvbuf(struct recv_buf *precvbuf, struct __queue *queue) argument 272 rtw_dequeue_recvbuf(struct __queue *queue) argument [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 15 * rvt_cq_enter - add a new entry to the completion queue 16 * @cq: completion queue 39 u_wc = cq->queue; 149 * rvt_create_cq - create a completion queue 183 * Allocate the completion queue entries and head/tail pointers. 254 cq->queue = u_wc; 270 * rvt_destroy_cq - destroy a completion queue 271 * @ibcq: the completion queue to destroy. 293 * rvt_req_notify_cq - change the notification type for a completion queue 294 * @ibcq: the completion queue [all...] |
/linux-master/drivers/accel/habanalabs/common/ |
H A D | irq.c | 29 * @ptr: the current ci or pi value of the completion queue 31 * Increment ptr by 1. If it reaches the number of completion queue 45 * @ptr: the current ci value of the event queue 47 * Increment ptr by 1. If it reaches the number of event queue 70 * job_finish - queue job finish work 74 * @cq: completion queue 80 struct hl_hw_queue *queue; local 83 queue = &hdev->kernel_queues[cq->hw_queue_id]; 84 job = queue->shadow_queue[hl_pi_2_offset(cs_seq)]; 88 atomic_inc(&queue 102 struct hl_hw_queue *queue; local [all...] |