/linux-master/drivers/mmc/core/ |
H A D | queue.c | 17 #include "queue.h" 140 struct request_queue *q = mq->queue; 382 mq->queue = disk->queue; 385 blk_queue_flag_set(QUEUE_FLAG_STABLE_WRITES, mq->queue); 386 blk_queue_rq_timeout(mq->queue, 60 * HZ); 388 blk_queue_flag_set(QUEUE_FLAG_NONROT, mq->queue); 389 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, mq->queue); 391 dma_set_max_seg_size(mmc_dev(host), queue_max_segment_size(mq->queue)); 400 mmc_crypto_setup_queue(mq->queue, hos [all...] |
H A D | Makefile | 18 mmc_block-objs := block.o queue.o
|
/linux-master/crypto/ |
H A D | algapi.c | 942 void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen) argument 944 INIT_LIST_HEAD(&queue->list); 945 queue->backlog = &queue->list; 946 queue->qlen = 0; 947 queue->max_qlen = max_qlen; 951 int crypto_enqueue_request(struct crypto_queue *queue, argument 956 if (unlikely(queue->qlen >= queue->max_qlen)) { 962 if (queue 974 crypto_enqueue_request_head(struct crypto_queue *queue, struct crypto_async_request *request) argument 985 crypto_dequeue_request(struct crypto_queue *queue) argument [all...] |
/linux-master/drivers/media/platform/sunxi/sun6i-csi/ |
H A D | sun6i_csi_capture.c | 464 list_for_each_entry(csi_buffer, &state->queue, list) { 470 INIT_LIST_HEAD(&state->queue); 483 if (list_empty(&state->queue)) 489 csi_buffer = list_first_entry(&state->queue, struct sun6i_csi_buffer, 551 static int sun6i_csi_capture_queue_setup(struct vb2_queue *queue, argument 557 struct sun6i_csi_device *csi_dev = vb2_get_drv_priv(queue); 600 list_add_tail(&csi_buffer->list, &state->queue); 604 static int sun6i_csi_capture_start_streaming(struct vb2_queue *queue, argument 607 struct sun6i_csi_device *csi_dev = vb2_get_drv_priv(queue); 638 static void sun6i_csi_capture_stop_streaming(struct vb2_queue *queue) argument 977 struct vb2_queue *queue = &capture->queue; local [all...] |
/linux-master/drivers/media/platform/renesas/vsp1/ |
H A D | vsp1_histo.c | 47 queue); 48 list_del(&buf->queue); 130 list_add_tail(&buf->queue, &histo->irqqueue); 147 /* Remove all buffers from the IRQ queue. */ 148 list_for_each_entry(buffer, &histo->irqqueue, queue) 445 if (f->index > 0 || f->type != histo->queue.type) 460 if (format->type != histo->queue.type) 562 /* ... and the buffers queue... */ 563 histo->queue.type = V4L2_BUF_TYPE_META_CAPTURE; 564 histo->queue [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | zcrypt_msgtype50.c | 189 * @zq: crypto queue pointer 252 * @zq: crypto queue pointer 358 AP_QID_CARD(zq->queue->qid), 359 AP_QID_QUEUE(zq->queue->qid), t80h->code); 361 __func__, AP_QID_CARD(zq->queue->qid), 362 AP_QID_QUEUE(zq->queue->qid), t80h->code); 363 ap_send_online_uevent(&zq->queue->ap_dev, zq->online); 391 AP_QID_CARD(zq->queue->qid), 392 AP_QID_QUEUE(zq->queue->qid), 396 __func__, AP_QID_CARD(zq->queue [all...] |
H A D | zcrypt_api.c | 568 static inline bool zcrypt_check_queue(struct ap_perms *perms, int queue) argument 570 return test_bit_inv(queue, perms->aqm) ? true : false; 578 if (!zq || !try_module_get(zq->queue->ap_dev.device.driver->owner)) 582 get_device(&zq->queue->ap_dev.device); 586 *pmod = zq->queue->ap_dev.device.driver->owner; 598 put_device(&zq->queue->ap_dev.device); 629 return zq->queue->total_request_count < 630 pref_zq->queue->total_request_count; 696 !ap_queue_usable(zq->queue)) 698 /* check if device node has admission for this queue */ 1274 int card, queue; local 1301 int card, queue; local 1322 zcrypt_device_status_ext(int card, int queue, struct zcrypt_device_status_ext *devstat) argument [all...] |
/linux-master/drivers/watchdog/ |
H A D | cpu5wdt.c | 51 int queue; member in struct:__anon47 71 if (cpu5wdt_device.queue && ticks) 95 if (!cpu5wdt_device.queue) { 96 cpu5wdt_device.queue = 1; 213 cpu5wdt_device.queue = 0; 252 if (cpu5wdt_device.queue) { 253 cpu5wdt_device.queue = 0;
|
H A D | rdc321x_wdt.c | 46 int queue; member in struct:__anon2921 74 if (rdc321x_wdt_device.queue && ticks) 93 if (!rdc321x_wdt_device.queue) { 94 rdc321x_wdt_device.queue = 1; 234 rdc321x_wdt_device.queue = 0; 262 if (rdc321x_wdt_device.queue) { 263 rdc321x_wdt_device.queue = 0;
|
/linux-master/block/ |
H A D | bsg.c | 24 struct request_queue *queue; member in struct:bsg_device 67 ret = bd->sg_io_fn(bd->queue, &hdr, open_for_write, 76 if (!blk_get_queue(to_bsg_device(inode)->queue)) 83 blk_put_queue(to_bsg_device(inode)->queue); 107 struct request_queue *q = bd->queue; 180 struct gendisk *disk = bd->queue->disk; 200 bd->queue = q;
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_srq.c | 68 rxe_dbg_srq(srq, "Unable to allocate queue\n"); 80 srq->rq.queue = q; 137 if (attr->srq_limit > srq->rq.queue->buf->index_mask) { 140 srq->rq.queue->buf->index_mask); 155 struct rxe_queue *q = srq->rq.queue; 186 srq->rq.queue = NULL; 197 if (srq->rq.queue) 198 rxe_queue_cleanup(srq->rq.queue);
|
H A D | rxe_qp.c | 201 qp->sq.queue = rxe_queue_init(rxe, &qp->sq.max_wr, wqe_size, 203 if (!qp->sq.queue) { 204 rxe_err_qp(qp, "Unable to allocate send queue\n"); 209 /* prepare info for caller to mmap send queue if user space qp */ 211 qp->sq.queue->buf, qp->sq.queue->buf_size, 212 &qp->sq.queue->ip); 228 vfree(qp->sq.queue->buf); 229 kfree(qp->sq.queue); 230 qp->sq.queue [all...] |
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_cldma.h | 94 int (*recv_skb)(struct cldma_queue *queue, struct sk_buff *skb); 129 void t7xx_cldma_set_recv_skb(struct cldma_queue *queue, 130 int (*recv_skb)(struct cldma_queue *queue, struct sk_buff *skb));
|
/linux-master/sound/core/seq/oss/ |
H A D | seq_oss_device.h | 75 int queue; /* sequencer queue number */ member in struct:seq_oss_devinfo 90 /* output queue */ 93 /* midi input queue */ 158 ev->queue = dp->queue;
|
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.c | 25 static void *hw_qpageit_get_inc(struct hw_queue *queue) argument 27 void *retvalue = hw_qeit_get(queue); 29 queue->current_q_offset += queue->pagesize; 30 if (queue->current_q_offset > queue->queue_length) { 31 queue->current_q_offset -= queue->pagesize; 40 static int hw_queue_ctor(struct hw_queue *queue, const u32 nr_of_pages, argument 52 queue 90 hw_queue_dtor(struct hw_queue *queue) argument [all...] |
/linux-master/drivers/usb/gadget/legacy/ |
H A D | raw_gadget.c | 54 static void raw_event_queue_init(struct raw_event_queue *queue) argument 56 spin_lock_init(&queue->lock); 57 sema_init(&queue->sema, 0); 58 queue->size = 0; 61 static int raw_event_queue_add(struct raw_event_queue *queue, argument 67 spin_lock_irqsave(&queue->lock, flags); 68 if (queue->size >= RAW_EVENT_QUEUE_SIZE) { 69 spin_unlock_irqrestore(&queue->lock, flags); 74 spin_unlock_irqrestore(&queue->lock, flags); 81 queue 88 raw_event_queue_fetch( struct raw_event_queue *queue) argument 120 raw_event_queue_destroy(struct raw_event_queue *queue) argument 185 struct raw_event_queue queue; member in struct:raw_dev [all...] |
/linux-master/sound/core/seq/ |
H A D | seq_clientmgr.c | 459 /* while data available in queue */ 580 bounce_ev.queue = SNDRV_SEQ_QUEUE_DIRECT; 600 * of the given queue. 604 int queue, int real_time) 608 q = queueptr(queue); 611 event->queue = queue; 742 update_timestamp_of_queue(event, subs->info.queue, 814 if (event->queue == SNDRV_SEQ_ADDRESS_SUBSCRIBERS || 825 * This function is called only from queue chec 603 update_timestamp_of_queue(struct snd_seq_event *event, int queue, int real_time) argument 1677 struct snd_seq_queue *queue; local 1706 struct snd_seq_queue *queue; local 1752 struct snd_seq_queue *queue; local [all...] |
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gem_submit.c | 32 struct msm_gpu_submitqueue *queue, uint32_t nr_bos, 57 ret = drm_sched_job_init(&submit->base, queue->entity, 1, queue); 66 submit->aspace = queue->ctx->aspace; 69 submit->queue = queue; 71 submit->ring = gpu->rb[queue->ring_nr]; 89 spin_lock(&submit->queue->idr_lock); 90 idr_remove(&submit->queue->fence_idr, submit->fence_id); 91 spin_unlock(&submit->queue 30 submit_create(struct drm_device *dev, struct msm_gpu *gpu, struct msm_gpu_submitqueue *queue, uint32_t nr_bos, uint32_t nr_cmds) argument 657 struct msm_gpu_submitqueue *queue; local [all...] |
/linux-master/drivers/media/platform/ti/omap3isp/ |
H A D | ispvideo.c | 318 * Video queue operations 321 static int isp_video_queue_setup(struct vb2_queue *queue, argument 325 struct isp_video_fh *vfh = vb2_get_drv_priv(queue); 350 * queue handler, which can't return an error, this check is just a best 371 * isp_video_buffer_queue - Add buffer to streaming queue 412 video->ops->queue(video, buffer); 450 static int isp_video_start_streaming(struct vb2_queue *queue, argument 453 struct isp_video_fh *vfh = vb2_get_drv_priv(queue); 494 * Remove the current video buffer from the DMA queue and fill its timestamp and 501 * The DMA queue i 1292 struct vb2_queue *queue; local [all...] |
/linux-master/drivers/crypto/cavium/zip/ |
H A D | zip_inflate.c | 133 u32 queue; local 140 /* Load inflate command to zip queue and ring the doorbell */ 141 queue = zip_load_instr(zip_cmd, zip_dev); 169 zip_update_cmd_bufs(zip_dev, queue); 173 zip_update_cmd_bufs(zip_dev, queue);
|
/linux-master/drivers/net/wireless/ralink/rt2x00/ |
H A D | rt2x00mmio.h | 88 * rt2x00mmio_flush_queue - Flush data queue 89 * @queue: Data queue to stop 95 void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop);
|
/linux-master/drivers/net/ethernet/netronome/nfp/abm/ |
H A D | main.h | 59 * @q_lvls: queue level control area 60 * @qm_stats: queue statistics symbol 61 * @q_stats: basic queue statistics (only in per-band case) 108 * @pdrop: number of hard drops due to queue limit 192 * @queue_base: id of base to host queue within PCIe (not QC idx) 256 unsigned int queue, u32 val); 260 unsigned int queue, enum nfp_abm_q_action act); 262 unsigned int band, unsigned int queue, 265 unsigned int band, unsigned int queue,
|
/linux-master/include/rdma/ |
H A D | rdmavt_cq.h | 34 * and completion queue entries as a single memory allocation so 44 * The completion queue structure. 55 struct rvt_cq_wc *queue; member in struct:rvt_cq
|
/linux-master/drivers/crypto/marvell/octeontx/ |
H A D | otx_cptvf_reqmgr.c | 271 struct otx_cpt_cmd_queue *queue; local 275 queue = &qinfo->queue[0]; 278 * therefore no locking is required for accessing instruction queue 280 ent = &queue->qhead->head[queue->idx * OTX_CPT_INST_SIZE]; 283 if (++queue->idx >= queue->qhead->size / 64) { 284 curr = queue->qhead; 286 if (list_is_last(&curr->nextchunk, &queue [all...] |
/linux-master/drivers/net/fddi/skfp/h/ |
H A D | hwmtm.h | 98 SMbuf *mb_free ; /* free queue */ 137 struct s_smt_tx_queue *tx_p ; /* pointer to the transmit queue */ 229 * txd *HWM_GET_TX_USED(smc,queue) 233 * number of used TxDs for the queue, specified by the index. 235 * para queue the number of the send queue: Can be specified by 238 * return number of used TxDs for this send queue 242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used 246 * txd *HWM_GET_CURR_TXD(smc,queue) [all...] |