Lines Matching defs:queue
84 struct pending_queue *queue;
86 for_each_pending_queue(pqinfo, queue, i) {
87 if (!queue->head)
90 /* free single queue */
91 kfree_sensitive((queue->head));
93 queue->front = 0;
94 queue->rear = 0;
108 struct pending_queue *queue = NULL;
113 for_each_pending_queue(pqinfo, queue, i) {
114 queue->head = kcalloc(qlen, sizeof(*queue->head), GFP_KERNEL);
115 if (!queue->head) {
120 queue->front = 0;
121 queue->rear = 0;
122 atomic64_set((&queue->pending_count), (0));
124 /* init queue spin lock */
125 spin_lock_init(&queue->lock);
161 dev_info(&pdev->dev, "Cleaning VQ pending queue (%u)\n",
170 struct command_queue *queue = NULL;
175 /* clean up for each queue */
177 queue = &cqinfo->queue[i];
178 if (hlist_empty(&cqinfo->queue[i].chead))
181 hlist_for_each_entry_safe(chunk, node, &cqinfo->queue[i].chead,
192 queue->nchunks = 0;
193 queue->idx = 0;
206 struct command_queue *queue = NULL;
217 /* per queue initialization */
224 queue = &cqinfo->queue[i];
225 INIT_HLIST_HEAD(&cqinfo->queue[i].chead);
239 i, queue->nchunks);
245 if (queue->nchunks == 0) {
247 &cqinfo->queue[i].chead);
254 queue->nchunks++;
262 /* Make the queue circular */
266 queue->qhead = curr;
267 spin_lock_init(&queue->lock);
300 dev_info(&pdev->dev, "Cleaning VQ command queue (%u)\n",
649 /* TODO: for now only one queue, so hard coded */
650 base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);